repo_name
stringlengths
6
97
path
stringlengths
3
341
text
stringlengths
8
1.02M
MaksGovor/FP-labs
lab2_OOP/src/main/scala/objsets/PostReader.scala
<reponame>MaksGovor/FP-labs package objsets import java.lang.StringBuffer object PostReader { object Parseposts { def regexParser(s: String): List[Map[String, Any]] = { // In real life. you would use an actual JSON library... val postRegex = """^\{ .*"user": "([^"]+)", "text": "([^"]+)", "likes": (...
MaksGovor/FP-labs
lab2_OOP/build.sbt
<gh_stars>0 name := "lab2_OOP" version := "0.1" scalaVersion := "2.13.6" libraryDependencies += "org.scalameta" %% "munit" % "0.4.3" % Test
MaksGovor/FP-labs
scalashop/src/main/scala/scalashop/Interfaces.scala
package scalashop // Interfaces used by the grading infrastructure. Do not change signatures // or your submission will fail with a NoSuchMethodError. trait HorizontalBoxBlurInterface { def blur(src: Img, dst: Img, from: Int, end: Int, radius: Int): Unit def parBlur(src: Img, dst: Img, numTasks: Int, radius: Int)...
MaksGovor/FP-labs
streams/src/main/scala/streams/StringParserTerrain.scala
<reponame>MaksGovor/FP-labs package streams /** * This component implements a parser to define terrains from a * graphical ASCII representation. * * When mixing in that component, a level can be defined by * defining the field `level` in the following form: * * val level = * """------ * |--ST-- *...
MaksGovor/FP-labs
quickcheck/src/main/scala/quickcheck/IndividualTask.scala
<filename>quickcheck/src/main/scala/quickcheck/IndividualTask.scala package quickcheck import org.scalacheck._ import Arbitrary._ import Gen._ import Prop._ import scala.annotation.tailrec trait IndividualTask { lazy val posGen: Gen[Int] = choose(2, 100) lazy val uncertaintyGen: Gen[Int] = oneOf(const(0), const(...
MaksGovor/FP-labs
scalashop/src/test/scala/scalashop/BlurSuite.scala
<gh_stars>0 package scalashop import java.util.concurrent._ import scala.collection._ import org.junit._ import org.junit.Assert.{assertArrayEquals, assertEquals} class BlurSuite { def boxBlurKernelBad(src: Img, x: Int, y: Int, radius: Int): RGBA = { val clampX = clamp(_, 0, src.width - 1) val clampY = cl...
MaksGovor/FP-labs
streams/build.sbt
name := "lab3_Collections" version := "0.1" scalaVersion := "2.13.6" libraryDependencies += "org.scalameta" %% "munit" % "0.7.27" % Test
MaksGovor/FP-labs
streams/src/test/scala/streams/BloxorzSuite.scala
<filename>streams/src/test/scala/streams/BloxorzSuite.scala package streams class BloxorzSuite extends munit.FunSuite { trait SolutionChecker extends GameDef with Solver with StringParserTerrain { /** * This method applies a list of moves `ls` to the block at position * `startPos`. This can be used to ...
MaksGovor/FP-labs
streams/src/test/scala/streams/IndividualTaskSuite.scala
package streams import streams.IndividualTask.{countExpression, toList} class IndividualTaskSuite extends munit.FunSuite { val inc: BigInt => BigInt = x => x + 1 val deInc: BigInt => BigInt = x => x - 1 val transform = List(inc, deInc) val range = (-250 to 25) val testList = range filter countExpression.is...
MaksGovor/FP-labs
effective-codecs/codecs/src/main/scala/codecs/Codecs.scala
package codecs /** * A data type modeling JSON values. * * For example, the `42` integer JSON value can be modeled as `Json.Num(42)` */ enum Json: /** * Try to decode this JSON value into a value of type `A` by using * the given decoder. * * Note that you have to explicitly fix `A` type param...
MaksGovor/FP-labs
lab_recursion/src/main/scala/recfun/Main.scala
<reponame>MaksGovor/FP-labs package recfun import scala.annotation.tailrec object Main { def main(args: Array[String]) { println("Pascal's Triangle") for (row <- 0 to 10) { for (col <- 0 to row) print(pascal(col, row) + " ") println() } println(calcExpression(6)) } /** * ...
MaksGovor/FP-labs
lab_recursion/build.sbt
<filename>lab_recursion/build.sbt<gh_stars>0 name := "lab_recursion" version := "0.1" scalaVersion := "2.13.6" libraryDependencies += "org.scalatest" %% "scalatest" % "3.1.0" % Test libraryDependencies += "junit" % "junit" % "4.10" % Test libraryDependencies += "org.scalatestplus" %% "junit-4-13" % "3.2.9.0" % "test...
MaksGovor/FP-labs
effective-codecs/codecs/build.sbt
scalaVersion := "3.1.0" scalacOptions ++= Seq("-deprecation") libraryDependencies ++= Seq( "org.typelevel" %% "jawn-parser" % "1.1.2", "org.scalameta" %% "munit" % "0.7.26" % Test, "org.scalameta" %% "munit-scalacheck" % "0.7.26" % Test )
MaksGovor/FP-labs
streams/src/main/scala/streams/IndividualTask.scala
<reponame>MaksGovor/FP-labs<gh_stars>0 package streams import scala.annotation.tailrec object IndividualTask { def main(args: Array[String]) { println("hello stub") val range = (-250 to 25) range.foreach(println) val list = toList(range, countExpression2) list.foreach(println) // val listMeth...
MaksGovor/FP-labs
scalashop/src/main/scala/scalashop/IndividualTask.scala
package scalashop import java.util.concurrent.{ConcurrentSkipListSet, TimeUnit} import scala.annotation.tailrec import scala.collection._ import scala.collection.parallel.CollectionConverters._ import scala.collection.parallel.ParSet import scala.concurrent.duration.{Duration, FiniteDuration} import scala.concurrent.{...
MaksGovor/FP-labs
quickcheck/src/main/scala/quickcheck/Heap.scala
package quickcheck trait IntHeap extends Heap { override type A = Int override def ord = scala.math.Ordering.Int } // http://www.brics.dk/RS/96/37/BRICS-RS-96-37.pdf // Figure 1, page 3 trait Heap { type H // type of a heap type A // type of an element def ord: Ordering[A] // ordering on elements def em...
devership16/INF553-YelpProject
src/data_preparation/Gov_Data_Las_Vegas_Prep.scala
<filename>src/data_preparation/Gov_Data_Las_Vegas_Prep.scala<gh_stars>1-10 import java.io.File import org.apache.hadoop.fs.FileUtil import org.apache.spark.{SparkConf, SparkContext} import org.apache.hadoop.conf.Configuration import org.apache.hadoop.fs._ import org.apache.spark.sql.SparkSession import org.apache.spar...
seglo/akka-projection
build.sbt
import akka.projections.Dependencies scalaVersion := "2.13.1" val commonSettings = Seq( organization := "com.lightbend.akka", scalacOptions ++= List( "-unchecked", "-deprecation", "-language:_", "-Xfatal-warnings", "-Ywarn-unused", "-encoding", "UTF-8"), javacOptions ++...
seglo/akka-projection
akka-projection-core/src/main/scala/akka/projection/Projection.scala
<reponame>seglo/akka-projection<gh_stars>1-10 /* * Copyright (C) 2020 Lightbend Inc. <https://www.lightbend.com> */ package akka.projection import akka.Done import akka.actor.ClassicActorSystemProvider import akka.annotation.ApiMayChange import scala.concurrent.{ ExecutionContext, Future } @ApiMayChange trait Pro...
seglo/akka-projection
project/Dependencies.scala
package akka.projections import sbt._ import sbt.Keys._ object Dependencies { object Versions { val akka = "2.6.4" val scalaTest = "3.1.1" } object Compile { val akkaStream = "com.typesafe.akka" %% "akka-stream" % Versions.akka } object Test { val scalaTest = "org.scalatest" %% "scalatest...
seglo/akka-projection
akka-projection-testkit/src/main/scala/akka/projection/testkit/ProjectionTestRunner.scala
<reponame>seglo/akka-projection<gh_stars>1-10 /* * Copyright (C) 2020 Lightbend Inc. <https://www.lightbend.com> */ package akka.projection.testkit import akka.actor.ClassicActorSystemProvider import akka.annotation.ApiMayChange import akka.projection.Projection import scala.concurrent.duration._ import scala.conc...
seglo/akka-projection
project/plugins.sbt
<reponame>seglo/akka-projection addSbtPlugin("org.scalameta" % "sbt-scalafmt" % "2.3.2") addSbtPlugin("de.heikoseeberger" % "sbt-header" % "5.4.0") addSbtPlugin("com.dwijnand" % "sbt-dynver" % "4.0.0") addSbtPlugin("com.lightbend.akka" % "sbt-paradox-akka" % "0.31") addSbtPlugin("com.lightbend" % "sbt-whitesource" % "...
Activiti/activiti-cloud-performance-tests
src/test/scala/simulation/BasicSimulation.scala
<reponame>Activiti/activiti-cloud-performance-tests package simulation import java.util.concurrent.TimeUnit import io.gatling.core.Predef._ import io.gatling.http.Predef._ class BasicSimulation extends Simulation { val httpConf = http .baseURL("http://" + System.getenv("GATEWAY_HOST")) val scn = scenario("...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson04_Spark_SQL/Lesson04_Exc03_GroupByExamples.scala
package org.pengfei.Lesson04_Spark_SQL import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.SparkSession import org.apache.spark.sql.functions.collect_list import org.apache.spark.sql.functions.collect_set object Lesson04_Exc03_GroupByExamples { def main(args:Array[String]):Unit={ Logger.getLogger("o...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson17_Analyze_Clinical_Data/Lesson17_Compare_Two_Columns.scala
package org.pengfei.Lesson17_Analyze_Clinical_Data import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.{DataFrame, SparkSession} import scala.collection.mutable.Map object Lesson17_Compare_Two_Columns { def main(args:Array[String]):Unit= { Logger.getLogger("org").setLevel(Level.OFF) Logger.g...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_5_3_3_Spark_ML_Exo1.scala
<filename>LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_5_3_3_Spark_ML_Exo1.scala<gh_stars>0 package org.pengfei.Lesson05_Spark_ML import org.apache.log4j.{Level, Logger} import org.apache.spark.ml.Pipeline import org.apache.spark.ml.classification.LogisticRegression import org.apache.spark.ml.eva...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson17_Analyze_Clinical_Data/Lesson17_WithColumn_When_Otherwise.scala
<reponame>pengfei99/Spark package org.pengfei.Lesson17_Analyze_Clinical_Data import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.{DataFrame, SparkSession} import org.apache.spark.sql.functions._ import org.apache.spark.sql.types.StringType //for 'when' object Lesson17_WithColumn_When_Otherwise { def...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson01_RDD/Lesson01_RDD_Basics.scala
<gh_stars>0 package org.pengfei.Lesson01_RDD import com.typesafe.config.ConfigFactory import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.SparkSession import org.apache.spark.storage.StorageLevel /** ************************************************************************************** * ************...
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/application/example/TaobaoSales.scala
package org.pengfei.spark.application.example import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.types.{IntegerType, StringType, StructField, StructType} import org.apache.spark.sql.{DataFrame, SparkSession} object TaobaoSales { def main(args:Array[String]): Unit ={ Logger.getLogger("org").setLeve...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson00_Spark_Core/Lesson00_Spark_Basic_Concept.scala
<filename>LearningSpark/src/main/java/org/pengfei/Lesson00_Spark_Core/Lesson00_Spark_Basic_Concept.scala package org.pengfei.Lesson00_Spark_Core import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.SparkSession object Lesson00_Spark_Basic_Concept { /**************************************************...
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/ml/classification/Word2Vec.scala
<gh_stars>0 package org.pengfei.spark.ml.classification import org.apache.log4j.{Level, Logger} import org.apache.spark.ml.feature.Word2Vec import org.apache.spark.sql.SparkSession object Word2Vec { def main(args:Array[String]): Unit ={ Logger.getLogger("org").setLevel(Level.OFF) Logger.getLogger("akka").setLevel...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/tmp/Test.scala
<reponame>pengfei99/Spark<gh_stars>0 package org.pengfei.tmp import java.util import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.{DataFrame, SparkSession} import org.apache.spark.sql.functions._ import org.apache.spark.sql.types.{BooleanType, IntegerType, StringType, StructField, StructType} import o...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_2_2_2_Anomaly_Detection_Algo.scala
package org.pengfei.Lesson05_Spark_ML object Lesson05_2_2_2_Anomaly_Detection_Algo { }
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson04_Spark_SQL/CustomArthMean.scala
<reponame>pengfei99/Spark package org.pengfei.Lesson04_Spark_SQL import org.apache.spark.sql.Row import org.apache.spark.sql.expressions.{MutableAggregationBuffer, UserDefinedAggregateFunction} import org.apache.spark.sql.types._ /* * This class extends UserDefinedAggregatedFunction to write custom arthmetic mean as...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson10_Spark_Application_ETL/Lesson10_1_record_deduplication.scala
package org.pengfei.Lesson10_Spark_Application_ETL import com.typesafe.config.ConfigFactory import org.apache.log4j.{Level, Logger} import org.apache.spark.rdd.RDD import org.apache.spark.sql.{DataFrame, SparkSession} import org.apache.spark.sql.functions._ import org.apache.spark.sql.Dataset object Lesson10_1_record...
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/formation/TweetsStat.scala
package org.pengfei.spark.formation import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.SparkSession import org.apache.spark.sql.functions.expr object TweetsStat { def main(args:Array[String]): Unit = { Logger.getLogger("org").setLevel(Level.OFF) Logger.getLogger("akka").setLevel(Level.OFF) ...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson10_Spark_Application_ETL/Pivot.scala
<gh_stars>0 package org.pengfei.Lesson10_Spark_Application_ETL import org.apache.spark.sql.{DataFrame, Dataset} import org.apache.spark.sql.functions.first object Pivot { def pivotSummary(df:DataFrame):DataFrame={ import df.sparkSession.implicits._ val schema= df.schema val ls:Dataset[(String,String,Do...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson21_Testing/Lesson21_1_Session_Creation.scala
<filename>LearningSpark/src/main/java/org/pengfei/Lesson21_Testing/Lesson21_1_Session_Creation.scala package org.pengfei.Lesson21_Testing object Lesson21_1_Session_Creation { }
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson04_Spark_SQL/Lesson04_5_Spark_DataSet.scala
package org.pengfei.Lesson04_Spark_SQL import java.util.Properties import com.typesafe.config.ConfigFactory import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.{Encoders, Row, SaveMode, SparkSession} import org.apache.spark.sql.types._ import org.apache.spark.storage.StorageLevel import org.pengfei.Le...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson12_Predicting_Forest_Cover_With_Decision_Trees/Lesson12_Predicting_Forest.scala
package org.pengfei.Lesson12_Predicting_Forest_Cover_With_Decision_Trees import com.typesafe.config.ConfigFactory import org.apache.log4j.{Level, Logger} import org.apache.spark.ml.{Pipeline, PipelineModel} import org.apache.spark.ml.classification.{DecisionTreeClassifier, RandomForestClassifier} import org.apache.spa...
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/application/example/IoTDeviceGeoIPDS.scala
package org.pengfei.spark.application.example import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.SparkSession object IoTDeviceGeoIPDS { def main(args:Array[String])={ Logger.getLogger("org").setLevel(Level.OFF) Logger.getLogger("akka").setLevel(Level.OFF) val spark = SparkSession.builder(). ...
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/application/example/TweeterPopularHashTags.scala
package org.pengfei.spark.application.example import org.apache.spark.streaming.twitter.TwitterUtils import org.apache.spark.streaming.{Seconds, StreamingContext} import org.apache.spark.{SparkConf, SparkContext} /** * A Spark Streaming application that receives tweets on certain * keywords from twitter data...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson01_RDD/Lesson01_RDDWithNumTypes.scala
package org.pengfei.Lesson01_RDD import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.SparkSession object Lesson01_RDDWithNumTypes { /* RDDs containing data elements of type Integer, Long, Float, or Double support a few additional actions that * are useful for statistical analysis.*/ def main(...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson06_Spark_Streaming/Lesson06_1_Spark_Streaming.scala
<filename>LearningSpark/src/main/java/org/pengfei/Lesson06_Spark_Streaming/Lesson06_1_Spark_Streaming.scala<gh_stars>0 package org.pengfei.Lesson06_Spark_Streaming import java.io.{BufferedReader, InputStream, InputStreamReader} import org.apache.log4j.receivers.net.SocketHubReceiver import org.apache.log4j.{Level, Lo...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_2_2_1_Clustering_Algo.scala
<reponame>pengfei99/Spark package org.pengfei.Lesson05_Spark_ML import org.apache.spark.sql.SparkSession object Lesson05_2_2_1_Clustering_Algo { /****************************************************************************************************************** * *****************************************5.2.2.1...
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/ml/classification/SupportVectorMachineRDD.scala
<gh_stars>0 package org.pengfei.spark.ml.classification import org.apache.log4j.{Level, Logger} import org.apache.spark.ml.feature.VectorAssembler import org.apache.spark.mllib.classification.SVMWithSGD import org.apache.spark.mllib.linalg.Vectors import org.apache.spark.mllib.regression.LabeledPoint import org.apache...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_2_2_3_dimensionality_reduction_Algo.scala
<filename>LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_2_2_3_dimensionality_reduction_Algo.scala package org.pengfei.Lesson05_Spark_ML object Lesson05_2_2_3_dimensionality_reduction_Algo { }
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_3_Model_Validation.scala
<reponame>pengfei99/Spark<gh_stars>0 package org.pengfei.Lesson05_Spark_ML import org.apache.spark.sql.SparkSession object Lesson05_3_Model_Validation { def main(args:Array[String])={ } /***********************************************AUC************************************************************/ def AUCExa...
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/ml/feature/transformation/StringFeatureTransformation.scala
package org.pengfei.spark.ml.feature.transformation import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.SparkSession import org.apache.spark.ml.feature._ import org.apache.spark.ml.linalg.Vectors object StringFeatureTransformation { /* * As most of the machine learning model does not deal with string/t...
pengfei99/Spark
Spark3/src/main/scala/org/pengfei/Lesson04_Spark_SQL/Lesson04_8_Spark_SQL_UDF.scala
package org.pengfei.Lesson04_Spark_SQL import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.{Column, SparkSession} import org.apache.spark.sql.functions.{col, lower, regexp_replace, udf} object Lesson04_8_Spark_SQL_UDF { def main(args: Array[String]) = { Logger.getLogger("org").setLevel(Level.OF...
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/basics/DataFrameWithArrayTypeAndMapType.scala
package org.pengfei.spark.basics import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.{Row, SparkSession} import org.apache.spark.sql.functions.split import org.apache.spark.sql.types._ /* * https://medium.com/@mrpowers/working-with-spark-arraytype-and-maptype-columns-4d85f3c8b2b3 * */ object DataFrameW...
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/streaming/flume/FlumeEventCount.scala
package org.pengfei.spark.streaming.flume import org.apache.spark.SparkConf import org.apache.spark.storage.StorageLevel import org.apache.spark.streaming._ import org.apache.spark.streaming.flume._ import org.apache.spark.util.IntParam object FlumeEventCount { def main(args: Array[String]) { //StreamingExample...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson14_Understanding_Wiki_With_Latent_Semantic_Analysis/Lesson14_Latent_Semantic_Analysis.scala
<filename>LearningSpark/src/main/java/org/pengfei/Lesson14_Understanding_Wiki_With_Latent_Semantic_Analysis/Lesson14_Latent_Semantic_Analysis.scala package org.pengfei.Lesson14_Understanding_Wiki_With_Latent_Semantic_Analysis import com.typesafe.config.ConfigFactory import org.apache.log4j.{Level, Logger} import org....
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson06_Spark_Streaming/Lesson06_4_Exo2.scala
<reponame>pengfei99/Spark package org.pengfei.Lesson06_Spark_Streaming import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.SparkSession /** In this exercise, we use a flume data source which output data to local socket 6666. * The objective of this exercise is to use scala streaming to read and proc...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_5_3_1_Spark_ML.scala
package org.pengfei.Lesson05_Spark_ML object Lesson05_5_3_1_Spark_ML { /********************************************************************************************************* * **********************************5.5.3 Spark Machine Learning pipelines API(Spark ML) *************************** * ***********...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson06_Spark_Streaming/Lesson06_4_Exo1.scala
<reponame>pengfei99/Spark package org.pengfei.Lesson06_Spark_Streaming import org.apache.spark._ import org.apache.spark.sql.SparkSession import org.apache.spark.streaming._ import org.apache.spark.streaming.twitter._ import twitter4j.Status object Lesson06_4_Exo1 { /* In this exercise, let’s develop a complete Spark...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/tmp/Realism_Data_Curation.scala
<reponame>pengfei99/Spark<filename>LearningSpark/src/main/java/org/pengfei/tmp/Realism_Data_Curation.scala package org.pengfei.Lesson17_Analyze_Clinical_Data import org.apache.log4j.{Level, Logger} import org.apache.spark.broadcast.Broadcast import org.apache.spark.sql.{DataFrame, SaveMode, SparkSession} import org.ap...
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/WordCount.scala
package org.pengfei.spark import org.apache.spark.{SparkConf, SparkContext} object WordCount { def main(args: Array[String]): Unit ={ val inputFile = "file:///tmp/word.txt" val conf = new SparkConf().setAppName("WordCount").setMaster("local") val sc = new SparkContext(conf) val textFile = sc.textFil...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson17_Analyze_Clinical_Data/Lesson17_Row_To_Column_To_Row.scala
package org.pengfei.Lesson17_Analyze_Clinical_Data import com.typesafe.config.ConfigFactory import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.{DataFrame, Dataset, SaveMode, SparkSession} import org.apache.spark.broadcast.Broadcast import org.apache.spark.sql.functions._ object Lesson17_Row_To_Column...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_5_2_2_MLlib_API_Models.scala
package org.pengfei.Lesson05_Spark_ML import java.util import org.apache.log4j.{Level, Logger} import org.apache.spark.SparkContext import org.apache.spark.ml.clustering.GaussianMixture import org.apache.spark.mllib.classification.SVMWithSGD import org.apache.spark.mllib.clustering.{GaussianMixtureModel, KMeans, KMea...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_5_2_1_MLlib_API_Data_Types.scala
<filename>LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_5_2_1_MLlib_API_Data_Types.scala package org.pengfei.Lesson05_Spark_ML import org.apache.log4j.{Level, Logger} import org.apache.spark.mllib.linalg.distributed._ import org.apache.spark.sql.SparkSession import org.apache.spark.mllib.linalg.{M...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson06_Spark_Streaming/SteamingSource.scala
<gh_stars>0 package org.pengfei.Lesson06_Spark_Streaming import java.io.PrintWriter import java.net.ServerSocket import scala.io.Source object SteamingSource { def index(n:Int)=scala.util.Random.nextInt(n) def main(args:Array[String]): Unit ={ // This object main takes three arguments, 1st is filePath, 2nd ...
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/application/example/Sanfransico911.scala
package org.pengfei.spark.application.example import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.SparkSession import org.apache.spark.sql.types._ //dependencies for timestamp functions (e.g. year, totimestamp) import org.apache.spark.sql.functions._ object Sanfransico911 { def main (args:Array[Strin...
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/ml/feature/select/FeatureSelection_ChiSqSelector.scala
<filename>WordCount/src/main/java/org/pengfei/spark/ml/feature/select/FeatureSelection_ChiSqSelector.scala package org.pengfei.spark.ml.feature.select import org.apache.log4j.{Level, Logger} import org.apache.spark.ml.feature.ChiSqSelector import org.apache.spark.ml.linalg.Vectors import org.apache.spark.sql.SparkSess...
pengfei99/Spark
common_utils/scala/data_validation.scala
/** * This function counts the null cell number * * @author <NAME> * @version 1.0 * @since 2020-01-27 * @param df source data frame * @param colName second column value to be merged * @return Long, It returns the number of null cell **/ def getNullCount(df: DataFrame, colName...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson06_Spark_Streaming/Lesson06_5_1_Spark_Structure_Streaming.scala
<reponame>pengfei99/Spark package org.pengfei.Lesson06_Spark_Streaming import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.SparkSession import org.apache.spark.sql.types.StructType object Lesson06_5_1_Spark_Structure_Streaming { /************************************************************************...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson16_Analyzing_Geospatial_Temporal_Data/Lesson16_Analyzing_Geo_Temporal_Data.scala
<filename>LearningSpark/src/main/java/org/pengfei/Lesson16_Analyzing_Geospatial_Temporal_Data/Lesson16_Analyzing_Geo_Temporal_Data.scala package org.pengfei.Lesson16_Analyzing_Geospatial_Temporal_Data import com.typesafe.config.ConfigFactory import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.SparkSess...
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/ml/classification/LogisticRegressionClassification.scala
<filename>WordCount/src/main/java/org/pengfei/spark/ml/classification/LogisticRegressionClassification.scala package org.pengfei.spark.ml.classification import org.apache.spark.ml.Pipeline import org.apache.spark.ml.classification.LogisticRegression import org.apache.spark.ml.feature._ import org.apache.spark.ml.linal...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_1_Spark_ML_Intro.scala
<reponame>pengfei99/Spark<filename>LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_1_Spark_ML_Intro.scala package org.pengfei.Lesson05_Spark_ML object Lesson05_1_Spark_ML_Intro { def main(args:Array[String])={ /***************************************************************************************...
pengfei99/Spark
WordCount/src/main/java/org/pengfei/test/Test.scala
<filename>WordCount/src/main/java/org/pengfei/test/Test.scala package org.pengfei.test import org.pengfei.spark.formation.TweetsStat.lineWordCount object Test { def main(args:Array[String]): Unit ={ val test="I'm <NAME>" print(lineWordCount(test)) } def lineWordCount(text: String): Long={ val wor...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson01_RDD/Lesson01_RDDWithKeyValuePair.scala
<reponame>pengfei99/Spark package org.pengfei.Lesson01_RDD import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.SparkSession /* * Key value pairs RDDs are a useful building block in many program * */ object Lesson01_RDDWithKeyValuePair { def main(args:Array[String]):Unit={ Logger.getLogger("org").set...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_2_1_1_Regression_Algo.scala
<gh_stars>0 package org.pengfei.Lesson05_Spark_ML import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.SparkSession object Lesson05_2_1_1_Regression_Algo { def main(args:Array[String])={ Logger.getLogger("org").setLevel(Level.OFF) Logger.getLogger("akka").setLevel(Level.OFF) val spark = ...
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/application/example/ClientSatisfait.scala
<gh_stars>0 package org.pengfei.spark.application.example import breeze.numerics.round //import org.apache.spark.sql.{Row, SQLContext, SparkSession} import org.apache.spark.sql.types._ import org.apache.spark.{SparkConf, SparkContext} object ClientSatisfait { def main(args:Array[String]): Unit ={ /*val inputFile =...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson21_Testing/TestGreeting.scala
<filename>LearningSpark/src/main/java/org/pengfei/Lesson21_Testing/TestGreeting.scala package org.pengfei.Lesson21_Testing class TestGreeting { }
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson04_Spark_SQL/Lesson04_Exc04_Parse_Apache_Access_Log.scala
package org.pengfei.Lesson04_Spark_SQL import com.typesafe.config.ConfigFactory import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.SparkSession import org.apache.spark.sql.functions._ object Lesson04_Exc04_Parse_Apache_Access_Log { def main(args: Array[String]): Unit = { Logger.getLogger("org"...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson04_Spark_SQL/Lesson04_7_Spark_SQL_Schema.scala
package org.pengfei.Lesson04_Spark_SQL import com.typesafe.config.ConfigFactory import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.{Row, SparkSession} import org.apache.spark.sql.types.{ArrayType, DataType, DataTypes, IntegerType, MapType, StringType, StructField, StructType} import org.apache.spark.s...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson14_Understanding_Wiki_With_Latent_Semantic_Analysis/Lesson14_Spark_Xml_Parsing.scala
<gh_stars>0 package org.pengfei.Lesson14_Understanding_Wiki_With_Latent_Semantic_Analysis import com.typesafe.config.ConfigFactory import org.apache.log4j.{Level, Logger} import org.apache.spark.ml.feature.{RegexTokenizer, StopWordsRemover, Tokenizer} import org.apache.spark.sql.{DataFrame, SparkSession} object Lesso...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson02_Spark_Jobs_And_Shared_Var/Lesson02_Spark_Jobs.scala
package org.pengfei.Lesson02_Spark_Jobs_And_Shared_Var import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.SparkSession object Lesson02_Spark_Jobs { def main(args:Array[String])={ Logger.getLogger("org").setLevel(Level.OFF) Logger.getLogger("akka").setLevel(Level.OFF) def spark=SparkSession.buil...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson04_Spark_SQL/Lesson04_6_Spark_SQL_Built_In_Functions.scala
<reponame>pengfei99/Spark package org.pengfei.Lesson04_Spark_SQL import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.SparkSession import org.apache.spark.sql.functions._ import org.apache.spark.sql.expressions.Window import org.pengfei.Lesson04_Spark_SQL.Lesson04_5_Spark_DataSet.{EmailArrayBody, EmailS...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson19_NLP/Lesson19_Keyword_extraction.scala
<filename>LearningSpark/src/main/java/org/pengfei/Lesson19_NLP/Lesson19_Keyword_extraction.scala<gh_stars>0 package org.pengfei.Lesson19_NLP import com.typesafe.config.ConfigFactory import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.functions._ import org.apache.spark.sql.types.StringType import org.a...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson04_Spark_SQL/Lesson04_Exc02_UDAF.scala
package org.pengfei.Lesson04_Spark_SQL import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.SparkSession object Lesson04_Exc02_UDAF { def main(args:Array[String]):Unit={ Logger.getLogger("org").setLevel(Level.OFF) Logger.getLogger("akka").setLevel(Level.OFF) val spark=SparkSession.builde...
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/ml/classification/LinearRegressionOnCalHousing.scala
package org.pengfei.spark.ml.classification import org.apache.log4j.{Level, Logger} import org.apache.spark.ml.Pipeline import org.apache.spark.ml.evaluation.RegressionEvaluator import org.apache.spark.ml.feature.VectorAssembler import org.apache.spark.sql.SparkSession import org.apache.spark.sql.types._ import org.ap...
pengfei99/Spark
common_utils/scala/data_io.scala
<filename>common_utils/scala/data_io.scala<gh_stars>0 /* This function write the input dataframe to the output file system*/ def WriteDataToDisk(df:DataFrame,outputPath:String,fileName:String): Unit ={ df.coalesce(1).write.mode(SaveMode.Overwrite) .option("header","true") .option("mapreduce.fileoutpu...
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/formation/USAFlight.scala
<reponame>pengfei99/Spark<filename>WordCount/src/main/java/org/pengfei/spark/formation/USAFlight.scala<gh_stars>0 package org.pengfei.spark.formation import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.SparkSession import org.apache.spark.sql.functions._ import org.apache.spark.sql.types._ object US...
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/SparkFileIO.scala
package org.pengfei.spark import org.apache.spark.{SparkConf, SparkContext} import scala.util.parsing.json.JSON object SparkFileIO { def main(args: Array[String]): Unit = { val conf = new SparkConf().setAppName("SparkFileIO").setMaster("local") val sc = new SparkContext(conf) val inputFile= "file:///...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson06_Spark_Streaming/Lesson06_3_Processing_Data_Stream.scala
<reponame>pengfei99/Spark package org.pengfei.Lesson06_Spark_Streaming import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.SparkSession import org.apache.spark.sql.execution.streaming.FileStreamSource.Timestamp import org.apache.spark.sql.functions.current_timestamp import org.apache.spark.streaming.ds...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson03_Spark_Application/Lesson03_1_Spark_Shell.scala
<gh_stars>0 package org.pengfei.Lesson03_Spark_Application object Lesson03_1_Spark_Shell { /***********************************************************************************/ }
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_2_1_2_Classification_Algo.scala
package org.pengfei.Lesson05_Spark_ML import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.SparkSession object Lesson05_2_1_2_Classification_Algo { /***************************************************************************************************** *************************************5.2.1.2 ...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson17_Analyze_Clinical_Data/Lesson17_Get_Specific_Row_Of_Column_Group.scala
package org.pengfei.Lesson17_Analyze_Clinical_Data import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.SparkSession import org.apache.spark.sql.functions._ import org.apache.spark.sql.expressions.Window object Lesson17_Get_Specific_Row_Of_Column_Group { /* In this section, we will show you how to g...
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/streaming/kafka/KafkaWordCount.scala
<gh_stars>0 package org.pengfei.spark.streaming.kafka import org.apache.spark.SparkConf import org.apache.spark.streaming._ import org.apache.spark.streaming.kafka._ object KafkaWordCount { def main(args:Array[String]){ //StreamingExamples.setStreamingLogLevels() val master = "spark://hadoop-nn.bioaster.org...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson06_Spark_Streaming/Lesson06_5_6_Structure_Streaming_Kafka_Integrations.scala
<reponame>pengfei99/Spark package org.pengfei.Lesson06_Spark_Streaming import org.apache.spark.sql.SparkSession object Lesson06_5_6_Structure_Streaming_Kafka_Integrations { def main(args:Array[String])={ } /******************************************************************************************************...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson10_Spark_Application_ETL/Lesson10_2_data_cleaning.scala
package org.pengfei.Lesson10_Spark_Application_ETL import com.typesafe.config.ConfigFactory import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.{DataFrame, Row, SparkSession} import org.apache.spark.sql.functions._ import org.apache.spark.sql.types.{DoubleType, IntegerType, StringType, StructField, Str...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson06_Spark_Streaming/Lesson06_5_5_Operations_On_Streaming_DF.scala
package org.pengfei.Lesson06_Spark_Streaming import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.expressions.scalalang.typed import org.apache.spark.sql.streaming.{OutputMode, Trigger} import org.apache.spark.sql.{DataFrame, Dataset, SparkSession} import org.apache.spark.sql.types._ import org.apache.s...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson04_Spark_SQL/Lesson04_1_Spark_SQL_Intro.scala
package org.pengfei.Lesson04_Spark_SQL import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.{DataFrame, SaveMode, SparkSession} object Lesson04_1_Spark_SQL_Intro { def main(args: Array[String]):Unit = { Logger.getLogger("org").setLevel(Level.OFF) Logger.getLogger("akka").setLevel(Level.OFF)...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson13_Anomaly_Detection/Lesson13_Anomaly_Detection_With_Streaming.scala
package org.pengfei.Lesson13_Anomaly_Detection import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.streaming.Trigger import org.apache.spark.sql.{Dataset, SparkSession} object Lesson13_Anomaly_Detection_With_Streaming { def main(args:Array[String]):Unit={ Logger.getLogger("org").setLevel(Level....
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/application/example/GPFSStats.scala
package org.pengfei.spark.application.example import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.SparkSession import org.apache.spark.sql.types._ import org.apache.spark.sql.functions._ import java.time.{Instant,ZoneId,ZonedDateTime} object GPFSStats { def main(args:Array[String]): Unit = { Logger....
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson03_Spark_Application/Lesson03_2_Spark_Application.scala
package org.pengfei.Lesson03_Spark_Application import org.apache.spark.sql.SparkSession object Lesson03_2_Spark_Application { def main(args: Array[String]) = { /** *****************************Spark API Entry Point: SparkSession ****************************/ /* Since spark 2.0+, SparkSession is the entry po...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson11_Spark_Application_Product_Recommendation/Lesson11_Recommendation.scala
package org.pengfei.Lesson11_Spark_Application_Product_Recommendation import com.typesafe.config.ConfigFactory import org.apache.log4j.{Level, Logger} import org.apache.spark.broadcast.Broadcast import org.apache.spark.sql.{DataFrame, Dataset, SparkSession} import org.apache.spark.sql.functions._ import org.apache.spa...
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/ml/feature/select/KeyWordSelection.scala
package org.pengfei.spark.ml.feature.select object KeyWordSelection { }
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/application/example/InvertedIndexShakespeare.scala
package org.pengfei.spark.application.example import java.io.File import org.apache.log4j.{Level, Logger} import org.apache.spark.sql._ import org.apache.spark.sql.types.{IntegerType, StringType, StructField, StructType} /* * * Inverted Index is mapping of content like text to the document in * which it can be foun...