repo_name stringlengths 6 97 | path stringlengths 3 341 | text stringlengths 8 1.02M |
|---|---|---|
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_6_Spark_Extra.scala | package org.pengfei.Lesson05_Spark_ML
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.{Row, SparkSession}
import org.apache.spark.sql.types.{IntegerType, StringType, StructField, StructType}
import org.apache.spark.sql.functions._
object Lesson05_6_Spark_Extra {
def main(args:Array[String])={
... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/basics/FileToRdd.scala | package org.pengfei.spark.basics
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.SparkSession
import org.apache.spark.sql.functions.typedLit
object FileToRdd {
def main(args:Array[String])={
Logger.getLogger("org").setLevel(Level.OFF)
Logger.getLogger("akka").setLevel(Level.OFF)
va... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson01_RDD/test.scala | package org.pengfei.Lesson01_RDD
object test {
def main(args:Array[String]):Unit={
val fileName="pliu.txt"
val badfn="pliu.txt.bkp"
println(getExtention(fileName))
println(getExtention(badfn))
}
def getExtention(fileName:String):String={
val index=fileName.lastIndexOf(".")
if(index>0)r... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/ml/feature/extraction/KeyWordExtraction.scala | package org.pengfei.spark.ml.feature.extraction
import org.apache.log4j.{Level, Logger}
import org.apache.spark.ml.feature.{HashingTF, IDF, Tokenizer}
import org.apache.spark.sql.SparkSession
/*
* In the following code segment, we start with a set of sentences.
* We split each sentence into words using Tokenizer.
* Fo... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/ml/data/preparation/NullValueEliminator.scala | <reponame>pengfei99/Spark
package org.pengfei.spark.ml.data.preparation
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.{DataFrame, SparkSession}
class NullValueEliminator {
def main(args:Array[String]): Unit ={
Logger.getLogger("org").setLevel(Level.OFF)
Logger.getLogger("akka").setLevel(Leve... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/basics/MultiDimentionalAgg.scala | <reponame>pengfei99/Spark
package org.pengfei.spark.basics
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.SparkSession
import org.apache.spark.sql.functions._
object MultiDimentionalAgg {
def main(args: Array[String])={
Logger.getLogger("org").setLevel(Level.OFF)
Logger.getLogger("akka... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson13_Anomaly_Detection/Lesson13_Anomaly_Detection.scala | <reponame>pengfei99/Spark<gh_stars>0
package org.pengfei.Lesson13_Anomaly_Detection
import com.typesafe.config.ConfigFactory
import org.apache.log4j.{Level, Logger}
import org.apache.spark.ml.{Pipeline, PipelineModel}
import org.apache.spark.ml.clustering.{KMeans, KMeansModel}
import org.apache.spark.ml.feature._
impo... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/ml/classification/SupportVectorMachineClassification.scala | <gh_stars>0
package org.pengfei.spark.ml.classification
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.SparkSession
import org.apache.spark.sql.types.{DoubleType, StringType, StructField, StructType}
import org.apache.spark.sql.functions.udf
import org.apache.spark.mllib.classification.SVMWithSGD
... |
pengfei99/Spark | common_utils/scala/data_transformation.scala | <filename>common_utils/scala/data_transformation.scala<gh_stars>0
/**
* This method transform multi rows of an object into columns, after the transformation, for each object we only have
* one row in the data frame. To make the transformation, this method cast all column to type string, as we don't do
* art... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson20_GPFS_Stats/Lesson20_GPFS_Stats.scala | <reponame>pengfei99/Spark<filename>LearningSpark/src/main/java/org/pengfei/Lesson20_GPFS_Stats/Lesson20_GPFS_Stats.scala<gh_stars>0
package org.pengfei.Lesson20_GPFS_Stats
import java.sql.Timestamp
import java.time._
import com.typesafe.config.ConfigFactory
import org.apache.log4j.{Level, Logger}
import org.apache.sp... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson17_Analyze_Clinical_Data/Lesson17_Analyze_Clinical_Data.scala | <reponame>pengfei99/Spark<gh_stars>0
package org.pengfei.Lesson17_Analyze_Clinical_Data
import com.typesafe.config.ConfigFactory
import org.apache.log4j.{Level, Logger}
import org.apache.spark.broadcast.Broadcast
import org.apache.spark.sql.{DataFrame, SaveMode, SparkSession}
import org.apache.spark.sql.functions._
im... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/ml/classification/DecisionTreeClassification.scala | package org.pengfei.spark.ml.classification
import org.apache.log4j.{Level, Logger}
import org.apache.spark.ml.Pipeline
import org.apache.spark.ml.classification.{DecisionTreeClassificationModel, DecisionTreeClassifier}
import org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
import org.apache.spark.ml.... |
pengfei99/Spark | WordCount/src/main/java/org/pengfei/spark/SparkHBaseIO.scala | package org.pengfei.spark
import org.apache.hadoop.conf.Configuration
import org.apache.hadoop.hbase.HBaseConfiguration
import org.apache.hadoop.hbase.client.{Put, Result}
import org.apache.hadoop.hbase.io.ImmutableBytesWritable
import org.apache.hadoop.hbase.mapreduce.{TableInputFormat, TableOutputFormat}
import org.... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson04_Spark_SQL/Lesson04_Exc01_yelp.scala | <reponame>pengfei99/Spark
package org.pengfei.Lesson04_Spark_SQL
import com.typesafe.config.ConfigFactory
import org.apache.log4j.{Level, Logger}
import org.apache.spark.sql.SparkSession
object Lesson04_Exc01_yelp {
def main(args:Array[String])={
Logger.getLogger("org").setLevel(Level.OFF)
Logger.getLogger("akka"... |
pengfei99/Spark | LearningSpark/src/main/java/org/pengfei/Lesson21_Testing/Greeting.scala | <filename>LearningSpark/src/main/java/org/pengfei/Lesson21_Testing/Greeting.scala
package org.pengfei.Lesson21_Testing
object Greeting {
}
|
comister/kayenta | kayenta-judge/src/main/scala/com/netflix/kayenta/judge/stats/DescriptiveStatistics.scala | <reponame>comister/kayenta<filename>kayenta-judge/src/main/scala/com/netflix/kayenta/judge/stats/DescriptiveStatistics.scala
/*
* Copyright 2017 Netflix, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License")
* you may not use this file except in compliance with the License.
* You may obtain a cop... |
comister/kayenta | kayenta-judge/src/test/scala/com/netflix/kayenta/judge/TestContextManagement.scala | /*
* Copyright 2017 Netflix, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License")
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to i... |
comister/kayenta | kayenta-judge/src/main/scala/com/netflix/kayenta/judge/detectors/KSigmaDetector.scala | <filename>kayenta-judge/src/main/scala/com/netflix/kayenta/judge/detectors/KSigmaDetector.scala
/*
* Copyright 2017 Netflix, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License")
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* h... |
comister/kayenta | kayenta-judge/src/main/scala/com/netflix/kayenta/judge/utils/RandomUtils.scala | package com.netflix.kayenta.judge.utils
import scala.util.Random
object RandomUtils {
private var random = new Random()
/**
* Initialize Random with the desired seed
*/
def init(seed: Int): Unit = {
random = new Random(seed)
}
/**
* Draw random samples from a normal (Gaussian) distributio... |
comister/kayenta | kayenta-judge/src/main/scala/com/netflix/kayenta/judge/utils/MapUtils.scala | <reponame>comister/kayenta
/*
* Copyright 2017 Netflix, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License")
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by ap... |
comister/kayenta | kayenta-judge/src/main/scala/com/netflix/kayenta/judge/detectors/IQRDetector.scala | <reponame>comister/kayenta
/*
* Copyright 2017 Netflix, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License")
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by ap... |
comister/kayenta | kayenta-judge/src/test/scala/com/netflix/kayenta/judge/StatisticSuite.scala | /*
* Copyright 2017 Netflix, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License")
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to i... |
comister/kayenta | kayenta-judge/src/test/scala/com/netflix/kayenta/judge/TransformSuite.scala | <filename>kayenta-judge/src/test/scala/com/netflix/kayenta/judge/TransformSuite.scala
/*
* Copyright 2017 Netflix, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License")
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.... |
comister/kayenta | kayenta-judge/src/main/scala/com/netflix/kayenta/judge/preprocessing/Validators.scala | <filename>kayenta-judge/src/main/scala/com/netflix/kayenta/judge/preprocessing/Validators.scala
/*
* Copyright 2017 Netflix, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License")
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* h... |
comister/kayenta | kayenta-judge/src/main/scala/com/netflix/kayenta/judge/classifiers/metric/MannWhitneyClassifier.scala | /*
* Copyright 2017 Netflix, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License")
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to i... |
comister/kayenta | kayenta-judge/src/main/scala/com/netflix/kayenta/judge/evaluation/BinaryClassificationEvaluator.scala | <filename>kayenta-judge/src/main/scala/com/netflix/kayenta/judge/evaluation/BinaryClassificationEvaluator.scala
/*
* Copyright 2017 Netflix, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License")
* you may not use this file except in compliance with the License.
* You may obtain a copy of the Lice... |
renhaocui/ensembleTopic | TMT/test2.scala | <reponame>renhaocui/ensembleTopic
import scalanlp.io._;
import scalanlp.stage._;
import scalanlp.stage.text._;
import scalanlp.text.tokenize._;
import scalanlp.pipes.Pipes.global._;
import edu.stanford.nlp.tmt.stage._;
import edu.stanford.nlp.tmt.model.lda._;
import edu.stanford.nlp.tmt.model.llda._;
val modelPath = ... |
renhaocui/ensembleTopic | TMT/train.scala | <gh_stars>1-10
import scalanlp.io._;
import scalanlp.stage._;
import scalanlp.stage.text._;
import scalanlp.text.tokenize._;
import scalanlp.pipes.Pipes.global._;
import edu.stanford.nlp.tmt.stage._;
import edu.stanford.nlp.tmt.model.lda._;
import edu.stanford.nlp.tmt.model.llda._;
val source = CSVFile("TMT\\LDAForma... |
f-loris/scio-idea-plugin | src/main/scala/com/spotify/scio/ScioInjector.scala | /*
* Copyright 2016 Spotify AB.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to ... |
guizmaii/BooleanDsl | src/main/scala/com/guizmaii/boolean/dsl/BooleanDslV1.scala | <reponame>guizmaii/BooleanDsl
package com.guizmaii.boolean.dsl
sealed trait BooleanDslV1 {
def unary_! : BooleanDslV1
final def &&(that: BooleanDslV1): BooleanDslV1 = BooleanDslV1.And(this, that)
final def ||(that: BooleanDslV1): BooleanDslV1 = BooleanDslV1.Or(this, that)
final def not: BooleanDslV1 ... |
guizmaii/BooleanDsl | build.sbt | name := "BooleanDsl"
version := "0.1"
scalaVersion := "2.13.6"
libraryDependencies ++= Seq(
"org.scalatest" %% "scalatest" % "3.2.3" % Test,
"org.scalatestplus" %% "scalacheck-1-14" % "3.2.2.0" % Test,
"org.scalacheck" %% "scalacheck" % "1.15.1" % Test
)
|
olka/stanoq | src/test/scala/org/stanoq/tests/crawler/ServiceNegativeSpec.scala | package org.stanoq.tests.crawler
import akka.http.scaladsl.server.ValidationRejection
import akka.http.scaladsl.testkit.ScalatestRouteTest
import org.scalatest._
import org.stanoq.crawler.CrawlerService
import org.stanoq.crawler.model.{ConfigProperties, CrawlerProtocols}
class ServiceNegativeSpec extends FlatSpec wit... |
olka/stanoq | src/test/scala/org/stanoq/load/tests/CrawlerLoadTest.scala | package org.stanoq.load.tests
import io.gatling.core.Predef._
import io.gatling.http.Predef._
import scala.concurrent.duration._
class CrawlerLoadTest extends Simulation {
val httpConf = http
.baseURL("http://localhost:9000")
.acceptHeader("application/json,application/xhtml+xml,application/xml;q=0.9,*/*... |
olka/stanoq | src/test/scala/org/stanoq/tests/crawler/CrawlerSpec.scala | package org.stanoq.tests.crawler
import org.scalatest._
import org.stanoq.crawler.Crawler
import spray.json._
import org.stanoq.crawler.model.{ConfigProperties, Node}
class CrawlerSpec extends FlatSpec with Matchers {
"Crawler" should "handle recursive page structure" in {
val crawler = new Crawler(ConfigPrope... |
olka/stanoq | build.sbt | name := "stanoq"
organization := "com.stanoq"
version := "1.0"
scalaVersion := "2.11.8"
scalacOptions := Seq("-unchecked", "-deprecation", "-encoding", "utf8")
enablePlugins(JavaAppPackaging)
enablePlugins(GatlingPlugin)
libraryDependencies ++= {
val akkaV = "2.4.16"
val akkaHttpV = "10.0.9"
val scalaTe... |
olka/stanoq | src/main/scala/org/stanoq/crawler/MongoHelper.scala | <gh_stars>1-10
package org.stanoq.crawler
import java.util.concurrent.TimeUnit
import com.typesafe.config.ConfigFactory
import org.mongodb.scala.{MongoClient, MongoCollection}
import org.stanoq.crawler.model._
import org.mongodb.scala.bson.codecs.Macros._
import org.mongodb.scala.bson.codecs.DEFAULT_CODEC_REGISTRY
im... |
olka/stanoq | src/test/scala/org/stanoq/load/tests/GatlingRunner.scala | <filename>src/test/scala/org/stanoq/load/tests/GatlingRunner.scala
package org.stanoq.load.tests
import io.gatling.app.Gatling
import io.gatling.core.config.GatlingPropertiesBuilder
object GatlingRunner extends App {
val props = new GatlingPropertiesBuilder
props.simulationClass(classOf[CrawlerLoadTest].getCanoni... |
olka/stanoq | src/main/scala/org/stanoq/crawler/Crawler.scala | <reponame>olka/stanoq
package org.stanoq.crawler
import java.util.Collections
import java.util.concurrent.{ConcurrentHashMap, TimeUnit}
import akka.actor.ActorSystem
import akka.event.Logging
import org.jsoup.nodes.Document
import org.jsoup.{Connection, HttpStatusException, Jsoup}
import org.stanoq.crawler.model.{Con... |
olka/stanoq | src/test/scala/org/stanoq/tests/crawler/ServiceSpec.scala | <filename>src/test/scala/org/stanoq/tests/crawler/ServiceSpec.scala
package org.stanoq.tests.crawler
import akka.event.NoLogging
import akka.http.scaladsl.model.ContentTypes._
import akka.http.scaladsl.model.StatusCodes._
import akka.http.scaladsl.testkit.{RouteTestTimeout, ScalatestRouteTest}
import org.scalatest._
i... |
olka/stanoq | src/main/scala/org/stanoq/version/VersionService.scala | <reponame>olka/stanoq<gh_stars>1-10
package org.stanoq.version
package org.stanoq.crawler
import akka.http.scaladsl.model.{HttpEntity, _}
import akka.http.scaladsl.server.Directives._
import com.typesafe.config.ConfigFactory
class VersionService {
val version = ConfigFactory.load().getString("stanoq.version")
... |
olka/stanoq | project/plugins.sbt | addSbtPlugin("com.eed3si9n" % "sbt-assembly" % "0.14.3")
addSbtPlugin("com.typesafe.sbt" % "sbt-native-packager" % "1.2.0-M7")
addSbtPlugin("org.scoverage" % "sbt-scoverage" % "1.5.0")
addSbtPlugin("io.gatling" % "gatling-sbt" % "2.2.0")
addSbtPlugin("org.scoverage" % "sbt-coveralls" % "1.1.0") |
olka/stanoq | src/test/scala/org/stanoq/load/tests/VersionLoadTest.scala | <reponame>olka/stanoq<filename>src/test/scala/org/stanoq/load/tests/VersionLoadTest.scala
package org.stanoq.load.tests
import io.gatling.core.Predef._
import io.gatling.http.Predef._
import scala.concurrent.duration._
class VersionLoadTest extends Simulation{
val httpConf = http.baseURL("http://stanoq.herokuapp.... |
olka/stanoq | src/main/scala/org/stanoq/CorsSupport.scala | package org.stanoq
import akka.http.scaladsl.model.HttpMethods._
import akka.http.scaladsl.model.{StatusCodes, HttpResponse}
import akka.http.scaladsl.model.headers._
import akka.http.scaladsl.server.Directives._
import akka.http.scaladsl.server.{Directive0, Route}
trait CorsSupport {
//this directive adds access ... |
olka/stanoq | src/test/scala/org/stanoq/tests/VersionSpec.scala | <reponame>olka/stanoq<filename>src/test/scala/org/stanoq/tests/VersionSpec.scala
package org.stanoq.tests
import akka.http.scaladsl.model.ContentTypes._
import akka.http.scaladsl.model.StatusCodes._
import akka.http.scaladsl.testkit.ScalatestRouteTest
import org.scalatest._
import org.stanoq.version.org.stanoq.crawler... |
olka/stanoq | src/main/scala/org/stanoq/RestController.scala | package org.stanoq
import akka.actor.ActorSystem
import akka.event.Logging
import akka.http.scaladsl.Http
import akka.http.scaladsl.server.Directives._
import akka.http.scaladsl.server.directives.DebuggingDirectives
import akka.stream.ActorMaterializer
import com.typesafe.config.ConfigFactory
import org.stanoq.crawler... |
olka/stanoq | src/main/scala/org/stanoq/auth/JwtAuth.scala | <filename>src/main/scala/org/stanoq/auth/JwtAuth.scala
package org.stanoq.auth
import pdi.jwt.{Jwt, JwtAlgorithm}
/**
* Jwt helper class
* @param payload @Json object converted to @String
* @param secret Combination of url and depth. This is UUID of crawling result
*/
case class JwtAuth(payload:String, secret... |
olka/stanoq | src/main/scala/org/stanoq/crawler/StreamService.scala | <reponame>olka/stanoq
package org.stanoq.crawler
import akka.actor.ActorSystem
import akka.http.scaladsl.common.{EntityStreamingSupport, JsonEntityStreamingSupport}
import akka.http.scaladsl.server.Directives._
import akka.stream.ThrottleMode
import akka.stream.impl.Stages.DefaultAttributes
import akka.stream.scaladsl... |
olka/stanoq | src/main/scala/org/stanoq/crawler/CrawlerService.scala | package org.stanoq.crawler
import akka.actor.ActorSystem
import akka.http.scaladsl.model._
import akka.http.scaladsl.server.Directives._
import org.stanoq.crawler.model._
import spray.json._
import scala.concurrent._
import scala.concurrent.Future
class CrawlerService() extends CrawlerProtocols {
implicit val blo... |
olka/stanoq | src/main/scala/org/stanoq/crawler/model/ConfigProperties.scala | <gh_stars>1-10
package org.stanoq.crawler.model
import java.awt.Color
import java.net.{URI, URL}
import akka.http.scaladsl.marshallers.sprayjson.SprayJsonSupport
import spray.json.{DefaultJsonProtocol, RootJsonFormat}
import scala.collection.mutable.Set
import scala.util.Try
case class ConfigProperties(url: String,... |
olka/stanoq | src/test/scala/org/stanoq/tests/crawler/AuthSpec.scala | package org.stanoq.tests.crawler
import org.scalatest._
import org.stanoq.auth.JwtAuth
import spray.json._
import org.stanoq.crawler.model.{ConfigProperties, CrawlerProtocols}
import pdi.jwt.exceptions.JwtValidationException
class AuthSpec extends FlatSpec with Matchers with CrawlerProtocols {
val config = ConfigP... |
Frugghi/TweetSpark | tweet-spark-scala/src/test/scala/com/tommasomadonia/spark/test/WordCountTests.scala | package com.tommasomadonia.spark.test
import com.tommasomadonia.spark.Words
import org.scalatest.{Matchers, GivenWhenThen, FlatSpec}
class WordCountTests extends FlatSpec with JSONSchemaSpec with GivenWhenThen with Matchers {
"Empty JSON" should "have no words" in {
Given("an empty DataFrame")
val dataFram... |
Frugghi/TweetSpark | tweet-spark-scala/src/main/scala/com/tommasomadonia/spark/TwitterAnalyzer.scala | package com.tommasomadonia.spark
import org.apache.hadoop.conf.Configuration
import org.apache.hadoop.fs.{FileSystem, Path}
import org.apache.spark.sql.SQLContext
import org.apache.spark.{SparkContext, SparkConf}
object TwitterAnalyzer {
def main(args: Array[String]) {
if (args.length < 1) {
System.err.p... |
Frugghi/TweetSpark | statsd-jvm-profiler/example/StatsDProfilerFlowListener.scala | package com.etsy.cascading.flow
import java.util.Properties
import cascading.flow.{Flow, FlowListener, FlowStep}
import org.apache.hadoop.mapred.JobConf
import scala.collection.JavaConversions._
/**
* Flow listener for setting up JobConf to enable statsd-jvm-profiler
*/
class StatsDProfilerFlowListener extends Fl... |
Frugghi/TweetSpark | tweet-spark-scala/src/test/scala/com/tommasomadonia/spark/test/SparkSpec.scala | package com.tommasomadonia.spark.test
import org.apache.log4j.{Level, Logger}
import org.apache.spark._
import org.scalatest._
trait SparkSpec extends BeforeAndAfterAll {
this: Suite =>
private val master = "local[2]"
private val appName = this.getClass.getSimpleName
private var _sparkContext: SparkContext ... |
Frugghi/TweetSpark | tweet-spark-scala/src/main/scala/com/tommasomadonia/spark/ActiveTweeters.scala | package com.tommasomadonia.spark
import org.apache.spark.sql.{DataFrame, SQLContext}
object ActiveTweeters {
def find(sqlContext: SQLContext, table: String): DataFrame = {
sqlContext.sql(s"""
|SELECT user.screen_name, COUNT(*) AS total_count
|FROM $table
... |
Frugghi/TweetSpark | tweet-spark-scala/src/main/scala/com/tommasomadonia/spark/TweetDataFrame.scala | <filename>tweet-spark-scala/src/main/scala/com/tommasomadonia/spark/TweetDataFrame.scala
package com.tommasomadonia.spark
import org.apache.spark.sql.{Row, DataFrame}
import org.apache.spark.sql.functions._
import scala.collection.mutable.ArrayBuffer
private case class _Indices(indices: Array[Long])
package object ... |
Frugghi/TweetSpark | tweet-spark-scala/src/test/scala/com/tommasomadonia/spark/test/SparkSQLSpec.scala | <reponame>Frugghi/TweetSpark
package com.tommasomadonia.spark.test
import org.apache.spark.sql.SQLContext
import org.scalatest.Suite
trait SparkSQLSpec extends SparkSpec {
this: Suite =>
private var _sqlContext: SQLContext = _
def sqlContext = _sqlContext
override def beforeAll(): Unit = {
super.before... |
Frugghi/TweetSpark | tweet-spark-scala/src/test/scala/com/tommasomadonia/spark/test/JSONSchemaSpec.scala | package com.tommasomadonia.spark.test
import org.apache.spark.sql.DataFrameReader
import org.apache.spark.sql.types.StructType
import org.scalatest.Suite
trait JSONSchemaSpec extends SparkSQLSpec {
this: Suite =>
private val jsonSchema = "test/schema.json"
private var _schema: StructType = _
private var _da... |
Frugghi/TweetSpark | tweet-spark-scala/src/main/scala/com/tommasomadonia/spark/Tweet.scala | package com.tommasomadonia.spark
import scala.collection.mutable.ArrayBuffer
case class Tweet(text: String, indices: Array[(Long, Long)]) {
def tokenize(): TraversableOnce[String] = this match {
case Tweet(tweet, indices) if (tweet == null || tweet.isEmpty) => Array[String]()
case Tweet(tweet, indices) => ... |
Frugghi/TweetSpark | tweet-spark-scala/src/main/scala/com/tommasomadonia/spark/Words.scala | package com.tommasomadonia.spark
import org.apache.spark.rdd.RDD
import org.apache.spark.sql.{DataFrame, Row}
import org.apache.spark.sql.functions._
import scala.collection.mutable.{ArrayBuffer, StringBuilder}
import com.github.nscala_time.time.Imports._
import com.tommasomadonia.spark.dataframe_extension._
priva... |
skygoo/octo | src/main/scala/org/seekloud/octo/ptcl/IceProtocol.scala | package org.seekloud.octo.ptcl
/**
* Created by sky
* Date on 2019/8/16
* Time at 17:48
*/
object IceProtocol {
case class CandidateInfo(
candidate: String,
sdpMid: String,
sdpMLineIndex: Int
)
}
|
skygoo/octo | src/main/scala/org/seekloud/octo/http/SessionBase.scala | package org.seekloud.octo.http
import akka.http.scaladsl.model._
import akka.http.scaladsl.server
import akka.http.scaladsl.server.Directives.{complete, extractRequestContext, onComplete, redirect, reject}
import akka.http.scaladsl.server._
import akka.http.scaladsl.server.directives.BasicDirectives
import org.seeklou... |
skygoo/octo | src/main/scala/org/seekloud/octo/http/HttpService.scala | package org.seekloud.octo.http
import akka.actor.{ActorSystem, Scheduler}
import akka.http.scaladsl.server.Directives._
import akka.http.scaladsl.server.Route
import akka.stream.Materializer
import akka.util.Timeout
import akka.actor.typed.scaladsl.AskPattern._
import scala.concurrent.Future
import scala.concurrent.E... |
skygoo/octo | src/main/scala/org/seekloud/octo/common/AppSettings.scala | // Copyright 2018 seekloud (https://github.com/seekloud)
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required b... |
skygoo/octo | src/main/scala/org/seekloud/octo/bridge/DtlsHandler.scala | <gh_stars>0
package org.seekloud.octo.bridge
import java.io.IOException
import javax.sdp.{MediaDescription, SdpException, SdpParseException, SessionDescription}
import org.ice4j.ice.{CandidatePair, Component, IceMediaStream, IceProcessingState}
import org.seekloud.octo.bridge.dtls.mock.Connect
import org.seekloud.oct... |
skygoo/octo | build.sbt | <filename>build.sbt
name := "octo"
version := "0.1"
scalaVersion := "2.12.8"
//resolvers += Resolver.mavenLocal
resolvers += "Jisti ORG Snapshots" at "https://github.com/jitsi/jitsi-maven-repository/raw/master/snapshots/"
//resolvers += "Jisti ORG Releases" at "https://github.com/jitsi/jitsi-maven-repository/raw/m... |
skygoo/octo | src/main/scala/org/seekloud/octo/ptcl/BrowserMsg.scala | <reponame>skygoo/octo
package org.seekloud.octo.ptcl
import org.seekloud.octo.ptcl.IceProtocol.CandidateInfo
/**
* Created by sky
* Date on 2019/8/16
* Time at 16:56
*/
object BrowserMsg {
object MsgId {
val PING = "PING" //前端定时发送
val PONG = "PONG" //后台回复
val Anchor_SDP_OFFER = "Anchor_SDP_OFFE... |
skygoo/octo | src/main/scala/org/seekloud/octo/bridge/IceHandler.scala | <filename>src/main/scala/org/seekloud/octo/bridge/IceHandler.scala<gh_stars>0
package org.seekloud.octo.bridge
import java.beans.{PropertyChangeEvent, PropertyChangeListener}
import java.io.IOException
import javax.sdp.{MediaDescription, SdpException, SessionDescription}
import org.ice4j.{Transport, TransportAddress}... |
skygoo/octo | src/main/scala/org/seekloud/octo/ptcl/EpInfo.scala | package org.seekloud.octo.ptcl
/**
* Created by sky
* Date on 2019/8/18
* Time at 14:24
* distinguish endpoint webSocketSession with info
*/
case class EpInfo(id:String)
|
skygoo/octo | src/main/scala/org/seekloud/octo/core/EndPointManager.scala | <reponame>skygoo/octo<filename>src/main/scala/org/seekloud/octo/core/EndPointManager.scala
package org.seekloud.octo.core
import akka.actor.typed.{ActorRef, Behavior}
import akka.actor.typed.scaladsl.{ActorContext, Behaviors, TimerScheduler}
import akka.http.scaladsl.model.ws.{Message, TextMessage}
import akka.stream.... |
skygoo/octo | project/Dependencies.scala | <filename>project/Dependencies.scala
import sbt._
/**
* User: Taoz
* Date: 6/13/2017
* Time: 9:38 PM
*/
object Dependencies {
val slickV = "3.2.3"
val akkaV = "2.5.22"
val akkaHttpV = "10.1.8"
val scalaXmlV = "1.1.0"
val circeVersion = "0.9.3"
val scalaJsDomV = "0.9.6"
val akkaSeq = Seq(
"... |
skygoo/octo | src/main/scala/org/seekloud/octo/core/EndPointWorker.scala | <reponame>skygoo/octo<gh_stars>0
package org.seekloud.octo.core
import java.io.FileInputStream
import akka.actor.typed.{ActorRef, Behavior}
import akka.actor.typed.scaladsl.{Behaviors, StashBuffer, TimerScheduler}
import akka.stream.OverflowStrategy
import akka.stream.scaladsl.Flow
import akka.stream.typed.scaladsl.{... |
skygoo/octo | src/main/scala/org/seekloud/octo/http/SocketService.scala | <gh_stars>0
package org.seekloud.octo.http
import akka.http.scaladsl.model.ws.Message
import akka.http.scaladsl.server.Directives._
import org.slf4j.LoggerFactory
import akka.actor.typed.scaladsl.AskPattern._
import akka.http.scaladsl.server.Route
import akka.stream.scaladsl.Flow
import scala.concurrent.Future
import... |
skygoo/octo | src/test/scala/org/seekloud/octo/Test.scala | package org.seekloud.octo
import java.beans.PropertyChangeEvent
import java.beans.PropertyChangeListener
import java.io.BufferedReader
import java.io.InputStreamReader
import java.net.DatagramSocket
import java.net.SocketAddress
import java.util
import org.ice4j.Transport
import org.ice4j.TransportAddress
import org.i... |
skygoo/octo | src/main/scala/org/seekloud/octo/Boot.scala | <filename>src/main/scala/org/seekloud/octo/Boot.scala
package org.seekloud.octo
import java.io.FileInputStream
import java.security.{KeyStore, SecureRandom}
import akka.actor.ActorSystem
import akka.actor.typed.ActorRef
import akka.actor.typed.scaladsl.adapter._
import akka.dispatch.MessageDispatcher
import akka.even... |
skygoo/octo | src/main/scala/org/seekloud/octo/ptcl/WebSocketSession.scala | package org.seekloud.octo.ptcl
import akka.actor.typed.ActorRef
import org.seekloud.octo.core.EndPointWorker
/**
* Created by sky
* Date on 2019/8/16
* Time at 15:16
*/
case class WebSocketSession(
id: String,
actor:ActorRef[EndPointWorker.Command],
... |
Axiometry/agario-client | src/main/scala/me/axiometry/agario/Agario.scala | <gh_stars>1-10
package me.axiometry.agario
import akka.io.IO
import akka.pattern.ask
import akka.actor.ActorSystem
import akka.util.Timeout
import spray.can.Http
import spray.http._
import HttpHeaders._
import HttpMethods._
import MediaTypes._
import spray.client.pipelining._
import scala.concurrent._
import scala.c... |
Axiometry/agario-client | src/main/scala/me/axiometry/agario/Entity.scala | <filename>src/main/scala/me/axiometry/agario/Entity.scala
package me.axiometry.agario
trait Entity {
def world: World
def x: Double
def x_=(x: Double): Unit
def y: Double
def y_=(y: Double): Unit
} |
Axiometry/agario-client | src/main/scala/me/axiometry/agario/net/DataView.scala | <reponame>Axiometry/agario-client
package me.axiometry.agario.net
import akka.util.ByteString
case class DataView(array: Array[Byte], offset: Int, length: Int)(implicit val byteOrder: ByteOrder = ByteOrder.BigEndian) {
import ByteOrder._
def getInt8(off: Int)(implicit byteOrder: ByteOrder = byteOrder): Byte =
... |
Axiometry/agario-client | src/main/scala/me/axiometry/agario/Client.scala | <filename>src/main/scala/me/axiometry/agario/Client.scala
package me.axiometry.agario
import me.axiometry.agario.net._
import akka.actor.{ ActorSystem, Props }
import grizzled.slf4j.Logging
import spray.can.Http
import spray.can.websocket.frame._
import spray.http._
import spray.httpx.RequestBuilding._
case class Cl... |
Axiometry/agario-client | src/main/scala/me/axiometry/agario/Server.scala | <filename>src/main/scala/me/axiometry/agario/Server.scala
package me.axiometry.agario
import akka.actor.ActorSystem
import scala.util._
case class ServerAddress(host: String, port: Int)
object ServerAddress {
private[this] val ServerPattern = "([^:]+):([0-9]+)".r
def unapply(string: String): Option[ServerAddress... |
Axiometry/agario-client | src/main/scala/me/axiometry/agario/World.scala | <reponame>Axiometry/agario-client<filename>src/main/scala/me/axiometry/agario/World.scala
package me.axiometry.agario
class World {
} |
Axiometry/agario-client | src/main/scala/me/axiometry/agario/net/WebSocketClient.scala | package me.axiometry.agario.net
import akka.actor.ActorSystem
import akka.io.IO
import spray.can.Http
import spray.can.server.UHttp
import spray.can.websocket._
import spray.can.websocket.frame._
import spray.http.HttpRequest
abstract class WebSocketClient(connect: Http.Connect, val upgradeRequest: HttpRequest)(impli... |
Axiometry/agario-client | src/main/scala/me/axiometry/agario/Carb.scala | <reponame>Axiometry/agario-client
package me.axiometry.agario
case class Carb(override val world: World) extends Entity {
override var x: Double = _
override var y: Double = _
} |
sequencer/rocket-dsp-utils | src/main/scala/freechips/rocketchip/jtag2mm/JtagFuzzer.scala | // SPDX-License-Identifier: Apache-2.0
package freechips.rocketchip.jtag2mm
import chisel3._
import chisel3.experimental._
import chisel3.util._
import chisel3.util.random.LFSR
class InvertedJtagIO extends Bundle {
// TRST (4.6) is optional and not currently implemented.
val TCK = Output(Bool())
val TMS = Out... |
sequencer/rocket-dsp-utils | build.sbt | <gh_stars>1-10
// SPDX-License-Identifier: Apache-2.0
git.remoteRepo := "<EMAIL>:ucb-bar/dsptools.git"
enablePlugins(SiteScaladocPlugin)
enablePlugins(GhpagesPlugin)
def scalacOptionsVersion(scalaVersion: String): Seq[String] = {
Seq() ++ {
// If we're building with Scala > 2.11, enable the compile option
... |
sequencer/rocket-dsp-utils | src/test/scala/freechips/rocketchip/jtag2mm/JtagFuzzerTester.scala | <filename>src/test/scala/freechips/rocketchip/jtag2mm/JtagFuzzerTester.scala
// SPDX-License-Identifier: Apache-2.0
package freechips.rocketchip.jtag2mm
import dsptools.DspTester
import org.scalatest.flatspec.AnyFlatSpec
import org.scalatest.matchers.should.Matchers
class JtagFuzzerTester(dut: JtagFuzzer) extends Ds... |
sequencer/rocket-dsp-utils | src/main/scala/package.scala | <filename>src/main/scala/package.scala
//val config = chipsalliance.rocketchip.config
|
sequencer/rocket-dsp-utils | src/main/scala/freechips/rocketchip/amba/axi4stream/SimpleSplitter.scala | package freechips.rocketchip.amba.axi4stream
import chisel3._
import freechips.rocketchip.amba.axi4stream.{AXI4StreamMasterPortParameters, AXI4StreamNexusNode, AXI4StreamSlavePortParameters}
import freechips.rocketchip.config.Parameters
import freechips.rocketchip.diplomacy._
class SimpleSplitter() extends LazyModule... |
sequencer/rocket-dsp-utils | src/main/scala/freechips/rocketchip/interrupts/Nodes.scala | package freechips.rocketchip.interrupts
import chisel3._
import chipsalliance.rocketchip.config.Parameters
import freechips.rocketchip.diplomacy._
object IntBundleBridgeImp extends BundleBridgeImp[Vec[Bool]]
case class IntToBundleBridgeNode(sinkParams: IntSinkPortParameters)(implicit valName: ValName)
extends MixedA... |
sequencer/rocket-dsp-utils | src/main/scala/freechips/rocketchip/jtag2mm/Utils.scala | // See ./LICENSE for license details.
package freechips.rocketchip.jtag2mm
import chisel3._
import chisel3.util._
// This code was taken from https://github.com/ucb-art/chisel-jtag/blob/master/src/main/scala/jtag/Utils.scala and adjusted to our design needs
/** Bundle representing a tristate pin.
*/
class Tristat... |
sequencer/rocket-dsp-utils | src/test/scala/dspblocks/DspBlockFirSpec.scala | <reponame>sequencer/rocket-dsp-utils<filename>src/test/scala/dspblocks/DspBlockFirSpec.scala
// SPDX-License-Identifier: Apache-2.0
package dspblocks
import chisel3._
import chisel3.iotesters._
import chisel3.util.Cat
import freechips.rocketchip.amba.axi4._
import freechips.rocketchip.amba.axi4stream._
import freechi... |
sequencer/rocket-dsp-utils | src/test/scala/dsptools/tester/MemMasterSpec.scala | package dsptools.tester
import chisel3._
import chisel3.iotesters.PeekPokeTester
import freechips.rocketchip.amba.apb._
import freechips.rocketchip.amba.axi4._
import freechips.rocketchip.config.Parameters
import freechips.rocketchip.diplomacy._
import freechips.rocketchip.interrupts._
import freechips.rocketchip.regm... |
tumsgis/veganmentor | src/test/scala/VeganMentorDbIntegrationTest.scala | <reponame>tumsgis/veganmentor
import java.time.LocalDateTime
import DataStructure.{Mentee, Mentor}
import junit.framework.TestCase
import org.junit.Assert._
import org.junit.{After, Before, Test}
class VeganMentorDbIntegrationTest extends TestCase {
@Before
override def setUp(): Unit = MongoDbRepo.dropAllPartic... |
tumsgis/veganmentor | src/main/scala/VeganMentor.scala | import DataStructure._
import scala.annotation.tailrec
object VeganMentor {
def saveInputFileToDb(csvFilePath: String): Unit = {
val (mentors, mentees) = Parser.parseInputFile(csvFilePath)
mentors.foreach(MongoDbRepo.saveParticipant)
mentees.foreach(MongoDbRepo.saveParticipant)
}
def process: Unit... |
tumsgis/veganmentor | src/test/scala/VeganMentorTest.scala | import junit.framework.TestCase
import org.junit.Assert._
import org.junit.{Before, Test}
import MongoDbRepo._
class VeganMentorTest extends TestCase{
@Before
override def setUp(): Unit = dropAllParticipants()
@Test
def testProcessing01(): Unit = {
val testFile = "src/test/files/Vegan_mentor_test_01.csv"... |
tumsgis/veganmentor | src/main/scala/Parser.scala | import java.io.File
import DataStructure.{Mentee, Mentor, SortedQueue, Participant}
import com.github.tototoshi.csv.CSVReader
object Parser {
def parseInputFile(filePath: String): (Seq[Mentor], Seq[Mentee]) = {
val reader = CSVReader.open(new File (filePath))
val participants: Seq[Participant] = reader.al... |
tumsgis/veganmentor | src/main/scala/Util.scala | import java.time.{LocalDateTime, ZoneId}
import java.time.format.DateTimeFormatter
import java.util.Date
object Util {
private val formatter = DateTimeFormatter.ofPattern("yyyy/MM/dd hh:mm:ss a z")
private val formatterWithSimplerHourFormat = DateTimeFormatter.ofPattern("yyyy/MM/dd h:mm:ss a z")
private def ge... |
tumsgis/veganmentor | src/main/scala/DataStructure.scala | import java.time.LocalDateTime
import com.mongodb.casbah.Imports
object DataStructure {
val mentorShipMap = Map("Mentor" -> true, "Lærlingur" -> false)
sealed trait Participant {
def timestamp: LocalDateTime
def email: String
def name: String
def note: String
def approvedTermsAndConditions: ... |
tumsgis/veganmentor | build.sbt | name := "VeganMentor"
version := "0.1"
scalaVersion := "2.12.4"
libraryDependencies += "com.github.tototoshi" %% "scala-csv" % "1.3.5"
libraryDependencies += "junit" % "junit" % "4.12" % Test
libraryDependencies += "com.google.apis" % "google-api-services-gmail" % "v1-rev75-1.23.0"
libraryDependencies += "com.goog... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.