repo_name stringlengths 6 97 | path stringlengths 3 341 | text stringlengths 8 1.02M |
|---|---|---|
MaksGovor/FP-labs | lab2_OOP/src/main/scala/objsets/PostReader.scala | <reponame>MaksGovor/FP-labs
package objsets
import java.lang.StringBuffer
object PostReader {
object Parseposts {
def regexParser(s: String): List[Map[String, Any]] = {
// In real life. you would use an actual JSON library...
val postRegex = """^\{ .*"user": "([^"]+)", "text": "([^"]+)", "likes": (... |
MaksGovor/FP-labs | lab2_OOP/build.sbt | <gh_stars>0
name := "lab2_OOP"
version := "0.1"
scalaVersion := "2.13.6"
libraryDependencies += "org.scalameta" %% "munit" % "0.4.3" % Test
|
MaksGovor/FP-labs | scalashop/src/main/scala/scalashop/Interfaces.scala | package scalashop
// Interfaces used by the grading infrastructure. Do not change signatures
// or your submission will fail with a NoSuchMethodError.
trait HorizontalBoxBlurInterface {
def blur(src: Img, dst: Img, from: Int, end: Int, radius: Int): Unit
def parBlur(src: Img, dst: Img, numTasks: Int, radius: Int)... |
MaksGovor/FP-labs | streams/src/main/scala/streams/StringParserTerrain.scala | <reponame>MaksGovor/FP-labs
package streams
/**
* This component implements a parser to define terrains from a
* graphical ASCII representation.
*
* When mixing in that component, a level can be defined by
* defining the field `level` in the following form:
*
* val level =
* """------
* |--ST--
*... |
MaksGovor/FP-labs | quickcheck/src/main/scala/quickcheck/IndividualTask.scala | <filename>quickcheck/src/main/scala/quickcheck/IndividualTask.scala
package quickcheck
import org.scalacheck._
import Arbitrary._
import Gen._
import Prop._
import scala.annotation.tailrec
trait IndividualTask {
lazy val posGen: Gen[Int] = choose(2, 100)
lazy val uncertaintyGen: Gen[Int] = oneOf(const(0), const(... |
MaksGovor/FP-labs | scalashop/src/test/scala/scalashop/BlurSuite.scala | <gh_stars>0
package scalashop
import java.util.concurrent._
import scala.collection._
import org.junit._
import org.junit.Assert.{assertArrayEquals, assertEquals}
class BlurSuite {
def boxBlurKernelBad(src: Img, x: Int, y: Int, radius: Int): RGBA = {
val clampX = clamp(_, 0, src.width - 1)
val clampY = cl... |
MaksGovor/FP-labs | streams/build.sbt | name := "lab3_Collections"
version := "0.1"
scalaVersion := "2.13.6"
libraryDependencies += "org.scalameta" %% "munit" % "0.7.27" % Test |
MaksGovor/FP-labs | streams/src/test/scala/streams/BloxorzSuite.scala | <filename>streams/src/test/scala/streams/BloxorzSuite.scala
package streams
class BloxorzSuite extends munit.FunSuite {
trait SolutionChecker extends GameDef with Solver with StringParserTerrain {
/**
* This method applies a list of moves `ls` to the block at position
* `startPos`. This can be used to ... |
MaksGovor/FP-labs | streams/src/test/scala/streams/IndividualTaskSuite.scala | package streams
import streams.IndividualTask.{countExpression, toList}
class IndividualTaskSuite extends munit.FunSuite {
val inc: BigInt => BigInt = x => x + 1
val deInc: BigInt => BigInt = x => x - 1
val transform = List(inc, deInc)
val range = (-250 to 25)
val testList = range filter countExpression.is... |
MaksGovor/FP-labs | effective-codecs/codecs/src/main/scala/codecs/Codecs.scala | package codecs
/**
* A data type modeling JSON values.
*
* For example, the `42` integer JSON value can be modeled as `Json.Num(42)`
*/
enum Json:
/**
* Try to decode this JSON value into a value of type `A` by using
* the given decoder.
*
* Note that you have to explicitly fix `A` type param... |
MaksGovor/FP-labs | lab_recursion/src/main/scala/recfun/Main.scala | <reponame>MaksGovor/FP-labs
package recfun
import scala.annotation.tailrec
object Main {
def main(args: Array[String]) {
println("Pascal's Triangle")
for (row <- 0 to 10) {
for (col <- 0 to row)
print(pascal(col, row) + " ")
println()
}
println(calcExpression(6))
}
/**
* ... |
MaksGovor/FP-labs | lab_recursion/build.sbt | <filename>lab_recursion/build.sbt<gh_stars>0
name := "lab_recursion"
version := "0.1"
scalaVersion := "2.13.6"
libraryDependencies += "org.scalatest" %% "scalatest" % "3.1.0" % Test
libraryDependencies += "junit" % "junit" % "4.10" % Test
libraryDependencies += "org.scalatestplus" %% "junit-4-13" % "3.2.9.0" % "test... |
MaksGovor/FP-labs | effective-codecs/codecs/build.sbt | scalaVersion := "3.1.0"
scalacOptions ++= Seq("-deprecation")
libraryDependencies ++= Seq(
"org.typelevel" %% "jawn-parser" % "1.1.2",
"org.scalameta" %% "munit" % "0.7.26" % Test,
"org.scalameta" %% "munit-scalacheck" % "0.7.26" % Test
)
|
MaksGovor/FP-labs | streams/src/main/scala/streams/IndividualTask.scala | <reponame>MaksGovor/FP-labs<gh_stars>0
package streams
import scala.annotation.tailrec
object IndividualTask {
def main(args: Array[String]) {
println("hello stub")
val range = (-250 to 25)
range.foreach(println)
val list = toList(range, countExpression2)
list.foreach(println)
// val listMeth... |
MaksGovor/FP-labs | scalashop/src/main/scala/scalashop/IndividualTask.scala | package scalashop
import java.util.concurrent.{ConcurrentSkipListSet, TimeUnit}
import scala.annotation.tailrec
import scala.collection._
import scala.collection.parallel.CollectionConverters._
import scala.collection.parallel.ParSet
import scala.concurrent.duration.{Duration, FiniteDuration}
import scala.concurrent.{... |
MaksGovor/FP-labs | quickcheck/src/main/scala/quickcheck/Heap.scala | package quickcheck
trait IntHeap extends Heap {
override type A = Int
override def ord = scala.math.Ordering.Int
}
// http://www.brics.dk/RS/96/37/BRICS-RS-96-37.pdf
// Figure 1, page 3
trait Heap {
type H // type of a heap
type A // type of an element
def ord: Ordering[A] // ordering on elements
def em... |
devership16/INF553-YelpProject | src/data_preparation/Gov_Data_Las_Vegas_Prep.scala | <filename>src/data_preparation/Gov_Data_Las_Vegas_Prep.scala<gh_stars>1-10
import java.io.File
import org.apache.hadoop.fs.FileUtil
import org.apache.spark.{SparkConf, SparkContext}
import org.apache.hadoop.conf.Configuration
import org.apache.hadoop.fs._
import org.apache.spark.sql.SparkSession
import org.apache.spar... |
seglo/akka-projection | build.sbt | import akka.projections.Dependencies
scalaVersion := "2.13.1"
val commonSettings = Seq(
organization := "com.lightbend.akka",
scalacOptions ++= List(
"-unchecked",
"-deprecation",
"-language:_",
"-Xfatal-warnings",
"-Ywarn-unused",
"-encoding",
"UTF-8"),
javacOptions ++... |
seglo/akka-projection | akka-projection-core/src/main/scala/akka/projection/Projection.scala | <reponame>seglo/akka-projection<gh_stars>1-10
/*
* Copyright (C) 2020 Lightbend Inc. <https://www.lightbend.com>
*/
package akka.projection
import akka.Done
import akka.actor.ClassicActorSystemProvider
import akka.annotation.ApiMayChange
import scala.concurrent.{ ExecutionContext, Future }
@ApiMayChange
trait Pro... |
seglo/akka-projection | project/Dependencies.scala | package akka.projections
import sbt._
import sbt.Keys._
object Dependencies {
object Versions {
val akka = "2.6.4"
val scalaTest = "3.1.1"
}
object Compile {
val akkaStream = "com.typesafe.akka" %% "akka-stream" % Versions.akka
}
object Test {
val scalaTest = "org.scalatest" %% "scalatest... |
seglo/akka-projection | akka-projection-testkit/src/main/scala/akka/projection/testkit/ProjectionTestRunner.scala | <reponame>seglo/akka-projection<gh_stars>1-10
/*
* Copyright (C) 2020 Lightbend Inc. <https://www.lightbend.com>
*/
package akka.projection.testkit
import akka.actor.ClassicActorSystemProvider
import akka.annotation.ApiMayChange
import akka.projection.Projection
import scala.concurrent.duration._
import scala.conc... |
seglo/akka-projection | project/plugins.sbt | <reponame>seglo/akka-projection
addSbtPlugin("org.scalameta" % "sbt-scalafmt" % "2.3.2")
addSbtPlugin("de.heikoseeberger" % "sbt-header" % "5.4.0")
addSbtPlugin("com.dwijnand" % "sbt-dynver" % "4.0.0")
addSbtPlugin("com.lightbend.akka" % "sbt-paradox-akka" % "0.31")
addSbtPlugin("com.lightbend" % "sbt-whitesource" % "... |
Activiti/activiti-cloud-performance-tests | src/test/scala/simulation/BasicSimulation.scala | <reponame>Activiti/activiti-cloud-performance-tests
package simulation
import java.util.concurrent.TimeUnit
import io.gatling.core.Predef._
import io.gatling.http.Predef._
class BasicSimulation extends Simulation {
val httpConf = http
.baseURL("http://" + System.getenv("GATEWAY_HOST"))
val scn = scenario("... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson04_Spark_SQL/Lesson04_Exc03_GroupByExamples.scala | package org.pengfei.Lesson04_Spark_SQL
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.SparkSession
import org.apache.spark.sql.functions.collect_list
import org.apache.spark.sql.functions.collect_set
object Lesson04_Exc03_GroupByExamples {
def main(args:Array[String]):Unit={
Logger.getLogger("o... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson17_Analyze_Clinical_Data/Lesson17_Compare_Two_Columns.scala | package org.pengfei.Lesson17_Analyze_Clinical_Data
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.{DataFrame, SparkSession}
import scala.collection.mutable.Map
object Lesson17_Compare_Two_Columns {
def main(args:Array[String]):Unit= {
Logger.getLogger("org").setLevel(Level.OFF)
Logger.g... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_5_3_3_Spark_ML_Exo1.scala | <filename>LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_5_3_3_Spark_ML_Exo1.scala<gh_stars>0
package org.pengfei.Lesson05_Spark_ML
import org.apache.log4j.{Level, Logger}
import org.apache.spark.ml.Pipeline
import org.apache.spark.ml.classification.LogisticRegression
import org.apache.spark.ml.eva... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson17_Analyze_Clinical_Data/Lesson17_WithColumn_When_Otherwise.scala | <reponame>pengfei99/Spark
package org.pengfei.Lesson17_Analyze_Clinical_Data
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.{DataFrame, SparkSession}
import org.apache.spark.sql.functions._
import org.apache.spark.sql.types.StringType //for 'when'
object Lesson17_WithColumn_When_Otherwise {
def... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson01_RDD/Lesson01_RDD_Basics.scala | <gh_stars>0
package org.pengfei.Lesson01_RDD
import com.typesafe.config.ConfigFactory
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.SparkSession
import org.apache.spark.storage.StorageLevel
/** **************************************************************************************
* ************... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/application/example/TaobaoSales.scala | package org.pengfei.spark.application.example
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.types.{IntegerType, StringType, StructField, StructType}
import org.apache.spark.sql.{DataFrame, SparkSession}
object TaobaoSales {
def main(args:Array[String]): Unit ={
Logger.getLogger("org").setLeve... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson00_Spark_Core/Lesson00_Spark_Basic_Concept.scala | <filename>LearningSpark/src/main/java/org/pengfei/Lesson00_Spark_Core/Lesson00_Spark_Basic_Concept.scala
package org.pengfei.Lesson00_Spark_Core
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.SparkSession
object Lesson00_Spark_Basic_Concept {
/**************************************************... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/ml/classification/Word2Vec.scala | <gh_stars>0
package org.pengfei.spark.ml.classification
import org.apache.log4j.{Level, Logger}
import org.apache.spark.ml.feature.Word2Vec
import org.apache.spark.sql.SparkSession
object Word2Vec {
def main(args:Array[String]): Unit ={
Logger.getLogger("org").setLevel(Level.OFF)
Logger.getLogger("akka").setLevel... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/tmp/Test.scala | <reponame>pengfei99/Spark<gh_stars>0
package org.pengfei.tmp
import java.util
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.{DataFrame, SparkSession}
import org.apache.spark.sql.functions._
import org.apache.spark.sql.types.{BooleanType, IntegerType, StringType, StructField, StructType}
import o... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_2_2_2_Anomaly_Detection_Algo.scala | package org.pengfei.Lesson05_Spark_ML
object Lesson05_2_2_2_Anomaly_Detection_Algo {
}
|
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson04_Spark_SQL/CustomArthMean.scala | <reponame>pengfei99/Spark
package org.pengfei.Lesson04_Spark_SQL
import org.apache.spark.sql.Row
import org.apache.spark.sql.expressions.{MutableAggregationBuffer, UserDefinedAggregateFunction}
import org.apache.spark.sql.types._
/*
* This class extends UserDefinedAggregatedFunction to write custom arthmetic mean as... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson10_Spark_Application_ETL/Lesson10_1_record_deduplication.scala | package org.pengfei.Lesson10_Spark_Application_ETL
import com.typesafe.config.ConfigFactory
import org.apache.log4j.{Level, Logger}
import org.apache.spark.rdd.RDD
import org.apache.spark.sql.{DataFrame, SparkSession}
import org.apache.spark.sql.functions._
import org.apache.spark.sql.Dataset
object Lesson10_1_record... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/formation/TweetsStat.scala | package org.pengfei.spark.formation
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.SparkSession
import org.apache.spark.sql.functions.expr
object TweetsStat {
def main(args:Array[String]): Unit = {
Logger.getLogger("org").setLevel(Level.OFF)
Logger.getLogger("akka").setLevel(Level.OFF)
... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson10_Spark_Application_ETL/Pivot.scala | <gh_stars>0
package org.pengfei.Lesson10_Spark_Application_ETL
import org.apache.spark.sql.{DataFrame, Dataset}
import org.apache.spark.sql.functions.first
object Pivot {
def pivotSummary(df:DataFrame):DataFrame={
import df.sparkSession.implicits._
val schema= df.schema
val ls:Dataset[(String,String,Do... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson21_Testing/Lesson21_1_Session_Creation.scala | <filename>LearningSpark/src/main/java/org/pengfei/Lesson21_Testing/Lesson21_1_Session_Creation.scala
package org.pengfei.Lesson21_Testing
object Lesson21_1_Session_Creation {
}
|
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson04_Spark_SQL/Lesson04_5_Spark_DataSet.scala | package org.pengfei.Lesson04_Spark_SQL
import java.util.Properties
import com.typesafe.config.ConfigFactory
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.{Encoders, Row, SaveMode, SparkSession}
import org.apache.spark.sql.types._
import org.apache.spark.storage.StorageLevel
import org.pengfei.Le... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson12_Predicting_Forest_Cover_With_Decision_Trees/Lesson12_Predicting_Forest.scala | package org.pengfei.Lesson12_Predicting_Forest_Cover_With_Decision_Trees
import com.typesafe.config.ConfigFactory
import org.apache.log4j.{Level, Logger}
import org.apache.spark.ml.{Pipeline, PipelineModel}
import org.apache.spark.ml.classification.{DecisionTreeClassifier, RandomForestClassifier}
import org.apache.spa... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/application/example/IoTDeviceGeoIPDS.scala | package org.pengfei.spark.application.example
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.SparkSession
object IoTDeviceGeoIPDS {
def main(args:Array[String])={
Logger.getLogger("org").setLevel(Level.OFF)
Logger.getLogger("akka").setLevel(Level.OFF)
val spark = SparkSession.builder().
... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/application/example/TweeterPopularHashTags.scala | package org.pengfei.spark.application.example
import org.apache.spark.streaming.twitter.TwitterUtils
import org.apache.spark.streaming.{Seconds, StreamingContext}
import org.apache.spark.{SparkConf, SparkContext}
/**
* A Spark Streaming application that receives tweets on certain
* keywords from twitter data... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson01_RDD/Lesson01_RDDWithNumTypes.scala | package org.pengfei.Lesson01_RDD
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.SparkSession
object Lesson01_RDDWithNumTypes {
/* RDDs containing data elements of type Integer, Long, Float, or Double support a few additional actions that
* are useful for statistical analysis.*/
def main(... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson06_Spark_Streaming/Lesson06_1_Spark_Streaming.scala | <filename>LearningSpark/src/main/java/org/pengfei/Lesson06_Spark_Streaming/Lesson06_1_Spark_Streaming.scala<gh_stars>0
package org.pengfei.Lesson06_Spark_Streaming
import java.io.{BufferedReader, InputStream, InputStreamReader}
import org.apache.log4j.receivers.net.SocketHubReceiver
import org.apache.log4j.{Level, Lo... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_2_2_1_Clustering_Algo.scala | <reponame>pengfei99/Spark
package org.pengfei.Lesson05_Spark_ML
import org.apache.spark.sql.SparkSession
object Lesson05_2_2_1_Clustering_Algo {
/******************************************************************************************************************
* *****************************************5.2.2.1... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/ml/classification/SupportVectorMachineRDD.scala | <gh_stars>0
package org.pengfei.spark.ml.classification
import org.apache.log4j.{Level, Logger}
import org.apache.spark.ml.feature.VectorAssembler
import org.apache.spark.mllib.classification.SVMWithSGD
import org.apache.spark.mllib.linalg.Vectors
import org.apache.spark.mllib.regression.LabeledPoint
import org.apache... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_2_2_3_dimensionality_reduction_Algo.scala | <filename>LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_2_2_3_dimensionality_reduction_Algo.scala
package org.pengfei.Lesson05_Spark_ML
object Lesson05_2_2_3_dimensionality_reduction_Algo {
}
|
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_3_Model_Validation.scala | <reponame>pengfei99/Spark<gh_stars>0
package org.pengfei.Lesson05_Spark_ML
import org.apache.spark.sql.SparkSession
object Lesson05_3_Model_Validation {
def main(args:Array[String])={
}
/***********************************************AUC************************************************************/
def AUCExa... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/ml/feature/transformation/StringFeatureTransformation.scala | package org.pengfei.spark.ml.feature.transformation
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.SparkSession
import org.apache.spark.ml.feature._
import org.apache.spark.ml.linalg.Vectors
object StringFeatureTransformation {
/*
* As most of the machine learning model does not deal with string/t... |
pengfei99/Spark | Spark3/src/main/scala/org/pengfei/Lesson04_Spark_SQL/Lesson04_8_Spark_SQL_UDF.scala | package org.pengfei.Lesson04_Spark_SQL
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.{Column, SparkSession}
import org.apache.spark.sql.functions.{col, lower, regexp_replace, udf}
object Lesson04_8_Spark_SQL_UDF {
def main(args: Array[String]) = {
Logger.getLogger("org").setLevel(Level.OF... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/basics/DataFrameWithArrayTypeAndMapType.scala | package org.pengfei.spark.basics
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.{Row, SparkSession}
import org.apache.spark.sql.functions.split
import org.apache.spark.sql.types._
/*
* https://medium.com/@mrpowers/working-with-spark-arraytype-and-maptype-columns-4d85f3c8b2b3
* */
object DataFrameW... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/streaming/flume/FlumeEventCount.scala | package org.pengfei.spark.streaming.flume
import org.apache.spark.SparkConf
import org.apache.spark.storage.StorageLevel
import org.apache.spark.streaming._
import org.apache.spark.streaming.flume._
import org.apache.spark.util.IntParam
object FlumeEventCount {
def main(args: Array[String]) {
//StreamingExample... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson14_Understanding_Wiki_With_Latent_Semantic_Analysis/Lesson14_Latent_Semantic_Analysis.scala | <filename>LearningSpark/src/main/java/org/pengfei/Lesson14_Understanding_Wiki_With_Latent_Semantic_Analysis/Lesson14_Latent_Semantic_Analysis.scala
package org.pengfei.Lesson14_Understanding_Wiki_With_Latent_Semantic_Analysis
import com.typesafe.config.ConfigFactory
import org.apache.log4j.{Level, Logger}
import org.... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson06_Spark_Streaming/Lesson06_4_Exo2.scala | <reponame>pengfei99/Spark
package org.pengfei.Lesson06_Spark_Streaming
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.SparkSession
/** In this exercise, we use a flume data source which output data to local socket 6666.
* The objective of this exercise is to use scala streaming to read and proc... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_5_3_1_Spark_ML.scala | package org.pengfei.Lesson05_Spark_ML
object Lesson05_5_3_1_Spark_ML {
/*********************************************************************************************************
* **********************************5.5.3 Spark Machine Learning pipelines API(Spark ML) ***************************
* ***********... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson06_Spark_Streaming/Lesson06_4_Exo1.scala | <reponame>pengfei99/Spark
package org.pengfei.Lesson06_Spark_Streaming
import org.apache.spark._
import org.apache.spark.sql.SparkSession
import org.apache.spark.streaming._
import org.apache.spark.streaming.twitter._
import twitter4j.Status
object Lesson06_4_Exo1 {
/* In this exercise, let’s develop a complete Spark... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/tmp/Realism_Data_Curation.scala | <reponame>pengfei99/Spark<filename>LearningSpark/src/main/java/org/pengfei/tmp/Realism_Data_Curation.scala
package org.pengfei.Lesson17_Analyze_Clinical_Data
import org.apache.log4j.{Level, Logger}
import org.apache.spark.broadcast.Broadcast
import org.apache.spark.sql.{DataFrame, SaveMode, SparkSession}
import org.ap... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/WordCount.scala | package org.pengfei.spark
import org.apache.spark.{SparkConf, SparkContext}
object WordCount {
def main(args: Array[String]): Unit ={
val inputFile = "file:///tmp/word.txt"
val conf = new SparkConf().setAppName("WordCount").setMaster("local")
val sc = new SparkContext(conf)
val textFile = sc.textFil... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson17_Analyze_Clinical_Data/Lesson17_Row_To_Column_To_Row.scala | package org.pengfei.Lesson17_Analyze_Clinical_Data
import com.typesafe.config.ConfigFactory
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.{DataFrame, Dataset, SaveMode, SparkSession}
import org.apache.spark.broadcast.Broadcast
import org.apache.spark.sql.functions._
object Lesson17_Row_To_Column... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_5_2_2_MLlib_API_Models.scala | package org.pengfei.Lesson05_Spark_ML
import java.util
import org.apache.log4j.{Level, Logger}
import org.apache.spark.SparkContext
import org.apache.spark.ml.clustering.GaussianMixture
import org.apache.spark.mllib.classification.SVMWithSGD
import org.apache.spark.mllib.clustering.{GaussianMixtureModel, KMeans, KMea... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_5_2_1_MLlib_API_Data_Types.scala | <filename>LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_5_2_1_MLlib_API_Data_Types.scala
package org.pengfei.Lesson05_Spark_ML
import org.apache.log4j.{Level, Logger}
import org.apache.spark.mllib.linalg.distributed._
import org.apache.spark.sql.SparkSession
import org.apache.spark.mllib.linalg.{M... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson06_Spark_Streaming/SteamingSource.scala | <gh_stars>0
package org.pengfei.Lesson06_Spark_Streaming
import java.io.PrintWriter
import java.net.ServerSocket
import scala.io.Source
object SteamingSource {
def index(n:Int)=scala.util.Random.nextInt(n)
def main(args:Array[String]): Unit ={
// This object main takes three arguments, 1st is filePath, 2nd ... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/application/example/Sanfransico911.scala | package org.pengfei.spark.application.example
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.SparkSession
import org.apache.spark.sql.types._
//dependencies for timestamp functions (e.g. year, totimestamp)
import org.apache.spark.sql.functions._
object Sanfransico911 {
def main (args:Array[Strin... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/ml/feature/select/FeatureSelection_ChiSqSelector.scala | <filename>WordCount/src/main/java/org/pengfei/spark/ml/feature/select/FeatureSelection_ChiSqSelector.scala
package org.pengfei.spark.ml.feature.select
import org.apache.log4j.{Level, Logger}
import org.apache.spark.ml.feature.ChiSqSelector
import org.apache.spark.ml.linalg.Vectors
import org.apache.spark.sql.SparkSess... |
pengfei99/Spark | common_utils/scala/data_validation.scala | /**
* This function counts the null cell number
*
* @author <NAME>
* @version 1.0
* @since 2020-01-27
* @param df source data frame
* @param colName second column value to be merged
* @return Long, It returns the number of null cell
**/
def getNullCount(df: DataFrame, colName... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson06_Spark_Streaming/Lesson06_5_1_Spark_Structure_Streaming.scala | <reponame>pengfei99/Spark
package org.pengfei.Lesson06_Spark_Streaming
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.SparkSession
import org.apache.spark.sql.types.StructType
object Lesson06_5_1_Spark_Structure_Streaming {
/************************************************************************... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson16_Analyzing_Geospatial_Temporal_Data/Lesson16_Analyzing_Geo_Temporal_Data.scala | <filename>LearningSpark/src/main/java/org/pengfei/Lesson16_Analyzing_Geospatial_Temporal_Data/Lesson16_Analyzing_Geo_Temporal_Data.scala
package org.pengfei.Lesson16_Analyzing_Geospatial_Temporal_Data
import com.typesafe.config.ConfigFactory
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.SparkSess... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/ml/classification/LogisticRegressionClassification.scala | <filename>WordCount/src/main/java/org/pengfei/spark/ml/classification/LogisticRegressionClassification.scala
package org.pengfei.spark.ml.classification
import org.apache.spark.ml.Pipeline
import org.apache.spark.ml.classification.LogisticRegression
import org.apache.spark.ml.feature._
import org.apache.spark.ml.linal... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_1_Spark_ML_Intro.scala | <reponame>pengfei99/Spark<filename>LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_1_Spark_ML_Intro.scala
package org.pengfei.Lesson05_Spark_ML
object Lesson05_1_Spark_ML_Intro {
def main(args:Array[String])={
/***************************************************************************************... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/test/Test.scala | <filename>WordCount/src/main/java/org/pengfei/test/Test.scala
package org.pengfei.test
import org.pengfei.spark.formation.TweetsStat.lineWordCount
object Test {
def main(args:Array[String]): Unit ={
val test="I'm <NAME>"
print(lineWordCount(test))
}
def lineWordCount(text: String): Long={
val wor... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson01_RDD/Lesson01_RDDWithKeyValuePair.scala | <reponame>pengfei99/Spark
package org.pengfei.Lesson01_RDD
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.SparkSession
/*
* Key value pairs RDDs are a useful building block in many program
* */
object Lesson01_RDDWithKeyValuePair {
def main(args:Array[String]):Unit={
Logger.getLogger("org").set... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_2_1_1_Regression_Algo.scala | <gh_stars>0
package org.pengfei.Lesson05_Spark_ML
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.SparkSession
object Lesson05_2_1_1_Regression_Algo {
def main(args:Array[String])={
Logger.getLogger("org").setLevel(Level.OFF)
Logger.getLogger("akka").setLevel(Level.OFF)
val spark = ... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/application/example/ClientSatisfait.scala | <gh_stars>0
package org.pengfei.spark.application.example
import breeze.numerics.round
//import org.apache.spark.sql.{Row, SQLContext, SparkSession}
import org.apache.spark.sql.types._
import org.apache.spark.{SparkConf, SparkContext}
object ClientSatisfait {
def main(args:Array[String]): Unit ={
/*val inputFile =... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson21_Testing/TestGreeting.scala | <filename>LearningSpark/src/main/java/org/pengfei/Lesson21_Testing/TestGreeting.scala
package org.pengfei.Lesson21_Testing
class TestGreeting {
}
|
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson04_Spark_SQL/Lesson04_Exc04_Parse_Apache_Access_Log.scala | package org.pengfei.Lesson04_Spark_SQL
import com.typesafe.config.ConfigFactory
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.SparkSession
import org.apache.spark.sql.functions._
object Lesson04_Exc04_Parse_Apache_Access_Log {
def main(args: Array[String]): Unit = {
Logger.getLogger("org"... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson04_Spark_SQL/Lesson04_7_Spark_SQL_Schema.scala | package org.pengfei.Lesson04_Spark_SQL
import com.typesafe.config.ConfigFactory
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.{Row, SparkSession}
import org.apache.spark.sql.types.{ArrayType, DataType, DataTypes, IntegerType, MapType, StringType, StructField, StructType}
import org.apache.spark.s... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson14_Understanding_Wiki_With_Latent_Semantic_Analysis/Lesson14_Spark_Xml_Parsing.scala | <gh_stars>0
package org.pengfei.Lesson14_Understanding_Wiki_With_Latent_Semantic_Analysis
import com.typesafe.config.ConfigFactory
import org.apache.log4j.{Level, Logger}
import org.apache.spark.ml.feature.{RegexTokenizer, StopWordsRemover, Tokenizer}
import org.apache.spark.sql.{DataFrame, SparkSession}
object Lesso... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson02_Spark_Jobs_And_Shared_Var/Lesson02_Spark_Jobs.scala | package org.pengfei.Lesson02_Spark_Jobs_And_Shared_Var
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.SparkSession
object Lesson02_Spark_Jobs {
def main(args:Array[String])={
Logger.getLogger("org").setLevel(Level.OFF)
Logger.getLogger("akka").setLevel(Level.OFF)
def spark=SparkSession.buil... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson04_Spark_SQL/Lesson04_6_Spark_SQL_Built_In_Functions.scala | <reponame>pengfei99/Spark
package org.pengfei.Lesson04_Spark_SQL
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.SparkSession
import org.apache.spark.sql.functions._
import org.apache.spark.sql.expressions.Window
import org.pengfei.Lesson04_Spark_SQL.Lesson04_5_Spark_DataSet.{EmailArrayBody, EmailS... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson19_NLP/Lesson19_Keyword_extraction.scala | <filename>LearningSpark/src/main/java/org/pengfei/Lesson19_NLP/Lesson19_Keyword_extraction.scala<gh_stars>0
package org.pengfei.Lesson19_NLP
import com.typesafe.config.ConfigFactory
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.functions._
import org.apache.spark.sql.types.StringType
import org.a... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson04_Spark_SQL/Lesson04_Exc02_UDAF.scala | package org.pengfei.Lesson04_Spark_SQL
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.SparkSession
object Lesson04_Exc02_UDAF {
def main(args:Array[String]):Unit={
Logger.getLogger("org").setLevel(Level.OFF)
Logger.getLogger("akka").setLevel(Level.OFF)
val spark=SparkSession.builde... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/ml/classification/LinearRegressionOnCalHousing.scala | package org.pengfei.spark.ml.classification
import org.apache.log4j.{Level, Logger}
import org.apache.spark.ml.Pipeline
import org.apache.spark.ml.evaluation.RegressionEvaluator
import org.apache.spark.ml.feature.VectorAssembler
import org.apache.spark.sql.SparkSession
import org.apache.spark.sql.types._
import org.ap... |
pengfei99/Spark | common_utils/scala/data_io.scala | <filename>common_utils/scala/data_io.scala<gh_stars>0
/* This function write the input dataframe to the output file system*/
def WriteDataToDisk(df:DataFrame,outputPath:String,fileName:String): Unit ={
df.coalesce(1).write.mode(SaveMode.Overwrite)
.option("header","true")
.option("mapreduce.fileoutpu... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/formation/USAFlight.scala | <reponame>pengfei99/Spark<filename>WordCount/src/main/java/org/pengfei/spark/formation/USAFlight.scala<gh_stars>0
package org.pengfei.spark.formation
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.SparkSession
import org.apache.spark.sql.functions._
import org.apache.spark.sql.types._
object US... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/SparkFileIO.scala | package org.pengfei.spark
import org.apache.spark.{SparkConf, SparkContext}
import scala.util.parsing.json.JSON
object SparkFileIO {
def main(args: Array[String]): Unit = {
val conf = new SparkConf().setAppName("SparkFileIO").setMaster("local")
val sc = new SparkContext(conf)
val inputFile= "file:///... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson06_Spark_Streaming/Lesson06_3_Processing_Data_Stream.scala | <reponame>pengfei99/Spark
package org.pengfei.Lesson06_Spark_Streaming
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.SparkSession
import org.apache.spark.sql.execution.streaming.FileStreamSource.Timestamp
import org.apache.spark.sql.functions.current_timestamp
import org.apache.spark.streaming.ds... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson03_Spark_Application/Lesson03_1_Spark_Shell.scala | <gh_stars>0
package org.pengfei.Lesson03_Spark_Application
object Lesson03_1_Spark_Shell {
/***********************************************************************************/
}
|
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_2_1_2_Classification_Algo.scala | package org.pengfei.Lesson05_Spark_ML
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.SparkSession
object Lesson05_2_1_2_Classification_Algo {
/*****************************************************************************************************
*************************************5.2.1.2 ... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson17_Analyze_Clinical_Data/Lesson17_Get_Specific_Row_Of_Column_Group.scala | package org.pengfei.Lesson17_Analyze_Clinical_Data
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.SparkSession
import org.apache.spark.sql.functions._
import org.apache.spark.sql.expressions.Window
object Lesson17_Get_Specific_Row_Of_Column_Group {
/* In this section, we will show you how to g... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/streaming/kafka/KafkaWordCount.scala | <gh_stars>0
package org.pengfei.spark.streaming.kafka
import org.apache.spark.SparkConf
import org.apache.spark.streaming._
import org.apache.spark.streaming.kafka._
object KafkaWordCount {
def main(args:Array[String]){
//StreamingExamples.setStreamingLogLevels()
val master = "spark://hadoop-nn.bioaster.org... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson06_Spark_Streaming/Lesson06_5_6_Structure_Streaming_Kafka_Integrations.scala | <reponame>pengfei99/Spark
package org.pengfei.Lesson06_Spark_Streaming
import org.apache.spark.sql.SparkSession
object Lesson06_5_6_Structure_Streaming_Kafka_Integrations {
def main(args:Array[String])={
}
/******************************************************************************************************... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson10_Spark_Application_ETL/Lesson10_2_data_cleaning.scala | package org.pengfei.Lesson10_Spark_Application_ETL
import com.typesafe.config.ConfigFactory
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.{DataFrame, Row, SparkSession}
import org.apache.spark.sql.functions._
import org.apache.spark.sql.types.{DoubleType, IntegerType, StringType, StructField, Str... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson06_Spark_Streaming/Lesson06_5_5_Operations_On_Streaming_DF.scala | package org.pengfei.Lesson06_Spark_Streaming
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.expressions.scalalang.typed
import org.apache.spark.sql.streaming.{OutputMode, Trigger}
import org.apache.spark.sql.{DataFrame, Dataset, SparkSession}
import org.apache.spark.sql.types._
import org.apache.s... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson04_Spark_SQL/Lesson04_1_Spark_SQL_Intro.scala | package org.pengfei.Lesson04_Spark_SQL
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.{DataFrame, SaveMode, SparkSession}
object Lesson04_1_Spark_SQL_Intro {
def main(args: Array[String]):Unit = {
Logger.getLogger("org").setLevel(Level.OFF)
Logger.getLogger("akka").setLevel(Level.OFF)... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson13_Anomaly_Detection/Lesson13_Anomaly_Detection_With_Streaming.scala | package org.pengfei.Lesson13_Anomaly_Detection
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.streaming.Trigger
import org.apache.spark.sql.{Dataset, SparkSession}
object Lesson13_Anomaly_Detection_With_Streaming {
def main(args:Array[String]):Unit={
Logger.getLogger("org").setLevel(Level.... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/application/example/GPFSStats.scala | package org.pengfei.spark.application.example
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.SparkSession
import org.apache.spark.sql.types._
import org.apache.spark.sql.functions._
import java.time.{Instant,ZoneId,ZonedDateTime}
object GPFSStats {
def main(args:Array[String]): Unit = {
Logger.... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson03_Spark_Application/Lesson03_2_Spark_Application.scala | package org.pengfei.Lesson03_Spark_Application
import org.apache.spark.sql.SparkSession
object Lesson03_2_Spark_Application {
def main(args: Array[String]) = {
/** *****************************Spark API Entry Point: SparkSession ****************************/
/* Since spark 2.0+, SparkSession is the entry po... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson11_Spark_Application_Product_Recommendation/Lesson11_Recommendation.scala | package org.pengfei.Lesson11_Spark_Application_Product_Recommendation
import com.typesafe.config.ConfigFactory
import org.apache.log4j.{Level, Logger}
import org.apache.spark.broadcast.Broadcast
import org.apache.spark.sql.{DataFrame, Dataset, SparkSession}
import org.apache.spark.sql.functions._
import org.apache.spa... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/ml/feature/select/KeyWordSelection.scala | package org.pengfei.spark.ml.feature.select
object KeyWordSelection {
}
|
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/application/example/InvertedIndexShakespeare.scala | package org.pengfei.spark.application.example
import java.io.File
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql._
import org.apache.spark.sql.types.{IntegerType, StringType, StructField, StructType}
/*
*
* Inverted Index is mapping of content like text to the document in
* which it can be foun... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.