repo_name
stringlengths
6
97
path
stringlengths
3
341
text
stringlengths
8
1.02M
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson05_Spark_ML/Lesson05_6_Spark_Extra.scala
package org.pengfei.Lesson05_Spark_ML import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.{Row, SparkSession} import org.apache.spark.sql.types.{IntegerType, StringType, StructField, StructType} import org.apache.spark.sql.functions._ object Lesson05_6_Spark_Extra { def main(args:Array[String])={ ...
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/basics/FileToRdd.scala
package org.pengfei.spark.basics import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.SparkSession import org.apache.spark.sql.functions.typedLit object FileToRdd { def main(args:Array[String])={ Logger.getLogger("org").setLevel(Level.OFF) Logger.getLogger("akka").setLevel(Level.OFF) va...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson01_RDD/test.scala
package org.pengfei.Lesson01_RDD object test { def main(args:Array[String]):Unit={ val fileName="pliu.txt" val badfn="pliu.txt.bkp" println(getExtention(fileName)) println(getExtention(badfn)) } def getExtention(fileName:String):String={ val index=fileName.lastIndexOf(".") if(index>0)r...
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/ml/feature/extraction/KeyWordExtraction.scala
package org.pengfei.spark.ml.feature.extraction import org.apache.log4j.{Level, Logger} import org.apache.spark.ml.feature.{HashingTF, IDF, Tokenizer} import org.apache.spark.sql.SparkSession /* * In the following code segment, we start with a set of sentences. * We split each sentence into words using Tokenizer. * Fo...
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/ml/data/preparation/NullValueEliminator.scala
<reponame>pengfei99/Spark package org.pengfei.spark.ml.data.preparation import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.{DataFrame, SparkSession} class NullValueEliminator { def main(args:Array[String]): Unit ={ Logger.getLogger("org").setLevel(Level.OFF) Logger.getLogger("akka").setLevel(Leve...
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/basics/MultiDimentionalAgg.scala
<reponame>pengfei99/Spark package org.pengfei.spark.basics import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.SparkSession import org.apache.spark.sql.functions._ object MultiDimentionalAgg { def main(args: Array[String])={ Logger.getLogger("org").setLevel(Level.OFF) Logger.getLogger("akka...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson13_Anomaly_Detection/Lesson13_Anomaly_Detection.scala
<reponame>pengfei99/Spark<gh_stars>0 package org.pengfei.Lesson13_Anomaly_Detection import com.typesafe.config.ConfigFactory import org.apache.log4j.{Level, Logger} import org.apache.spark.ml.{Pipeline, PipelineModel} import org.apache.spark.ml.clustering.{KMeans, KMeansModel} import org.apache.spark.ml.feature._ impo...
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/ml/classification/SupportVectorMachineClassification.scala
<gh_stars>0 package org.pengfei.spark.ml.classification import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.SparkSession import org.apache.spark.sql.types.{DoubleType, StringType, StructField, StructType} import org.apache.spark.sql.functions.udf import org.apache.spark.mllib.classification.SVMWithSGD ...
pengfei99/Spark
common_utils/scala/data_transformation.scala
<filename>common_utils/scala/data_transformation.scala<gh_stars>0 /** * This method transform multi rows of an object into columns, after the transformation, for each object we only have * one row in the data frame. To make the transformation, this method cast all column to type string, as we don't do * art...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson20_GPFS_Stats/Lesson20_GPFS_Stats.scala
<reponame>pengfei99/Spark<filename>LearningSpark/src/main/java/org/pengfei/Lesson20_GPFS_Stats/Lesson20_GPFS_Stats.scala<gh_stars>0 package org.pengfei.Lesson20_GPFS_Stats import java.sql.Timestamp import java.time._ import com.typesafe.config.ConfigFactory import org.apache.log4j.{Level, Logger} import org.apache.sp...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson17_Analyze_Clinical_Data/Lesson17_Analyze_Clinical_Data.scala
<reponame>pengfei99/Spark<gh_stars>0 package org.pengfei.Lesson17_Analyze_Clinical_Data import com.typesafe.config.ConfigFactory import org.apache.log4j.{Level, Logger} import org.apache.spark.broadcast.Broadcast import org.apache.spark.sql.{DataFrame, SaveMode, SparkSession} import org.apache.spark.sql.functions._ im...
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/ml/classification/DecisionTreeClassification.scala
package org.pengfei.spark.ml.classification import org.apache.log4j.{Level, Logger} import org.apache.spark.ml.Pipeline import org.apache.spark.ml.classification.{DecisionTreeClassificationModel, DecisionTreeClassifier} import org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator import org.apache.spark.ml....
pengfei99/Spark
WordCount/src/main/java/org/pengfei/spark/SparkHBaseIO.scala
package org.pengfei.spark import org.apache.hadoop.conf.Configuration import org.apache.hadoop.hbase.HBaseConfiguration import org.apache.hadoop.hbase.client.{Put, Result} import org.apache.hadoop.hbase.io.ImmutableBytesWritable import org.apache.hadoop.hbase.mapreduce.{TableInputFormat, TableOutputFormat} import org....
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson04_Spark_SQL/Lesson04_Exc01_yelp.scala
<reponame>pengfei99/Spark package org.pengfei.Lesson04_Spark_SQL import com.typesafe.config.ConfigFactory import org.apache.log4j.{Level, Logger} import org.apache.spark.sql.SparkSession object Lesson04_Exc01_yelp { def main(args:Array[String])={ Logger.getLogger("org").setLevel(Level.OFF) Logger.getLogger("akka"...
pengfei99/Spark
LearningSpark/src/main/java/org/pengfei/Lesson21_Testing/Greeting.scala
<filename>LearningSpark/src/main/java/org/pengfei/Lesson21_Testing/Greeting.scala package org.pengfei.Lesson21_Testing object Greeting { }
comister/kayenta
kayenta-judge/src/main/scala/com/netflix/kayenta/judge/stats/DescriptiveStatistics.scala
<reponame>comister/kayenta<filename>kayenta-judge/src/main/scala/com/netflix/kayenta/judge/stats/DescriptiveStatistics.scala /* * Copyright 2017 Netflix, Inc. * * Licensed under the Apache License, Version 2.0 (the "License") * you may not use this file except in compliance with the License. * You may obtain a cop...
comister/kayenta
kayenta-judge/src/test/scala/com/netflix/kayenta/judge/TestContextManagement.scala
/* * Copyright 2017 Netflix, Inc. * * Licensed under the Apache License, Version 2.0 (the "License") * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to i...
comister/kayenta
kayenta-judge/src/main/scala/com/netflix/kayenta/judge/detectors/KSigmaDetector.scala
<filename>kayenta-judge/src/main/scala/com/netflix/kayenta/judge/detectors/KSigmaDetector.scala /* * Copyright 2017 Netflix, Inc. * * Licensed under the Apache License, Version 2.0 (the "License") * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * h...
comister/kayenta
kayenta-judge/src/main/scala/com/netflix/kayenta/judge/utils/RandomUtils.scala
package com.netflix.kayenta.judge.utils import scala.util.Random object RandomUtils { private var random = new Random() /** * Initialize Random with the desired seed */ def init(seed: Int): Unit = { random = new Random(seed) } /** * Draw random samples from a normal (Gaussian) distributio...
comister/kayenta
kayenta-judge/src/main/scala/com/netflix/kayenta/judge/utils/MapUtils.scala
<reponame>comister/kayenta /* * Copyright 2017 Netflix, Inc. * * Licensed under the Apache License, Version 2.0 (the "License") * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by ap...
comister/kayenta
kayenta-judge/src/main/scala/com/netflix/kayenta/judge/detectors/IQRDetector.scala
<reponame>comister/kayenta /* * Copyright 2017 Netflix, Inc. * * Licensed under the Apache License, Version 2.0 (the "License") * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by ap...
comister/kayenta
kayenta-judge/src/test/scala/com/netflix/kayenta/judge/StatisticSuite.scala
/* * Copyright 2017 Netflix, Inc. * * Licensed under the Apache License, Version 2.0 (the "License") * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to i...
comister/kayenta
kayenta-judge/src/test/scala/com/netflix/kayenta/judge/TransformSuite.scala
<filename>kayenta-judge/src/test/scala/com/netflix/kayenta/judge/TransformSuite.scala /* * Copyright 2017 Netflix, Inc. * * Licensed under the Apache License, Version 2.0 (the "License") * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www....
comister/kayenta
kayenta-judge/src/main/scala/com/netflix/kayenta/judge/preprocessing/Validators.scala
<filename>kayenta-judge/src/main/scala/com/netflix/kayenta/judge/preprocessing/Validators.scala /* * Copyright 2017 Netflix, Inc. * * Licensed under the Apache License, Version 2.0 (the "License") * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * h...
comister/kayenta
kayenta-judge/src/main/scala/com/netflix/kayenta/judge/classifiers/metric/MannWhitneyClassifier.scala
/* * Copyright 2017 Netflix, Inc. * * Licensed under the Apache License, Version 2.0 (the "License") * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to i...
comister/kayenta
kayenta-judge/src/main/scala/com/netflix/kayenta/judge/evaluation/BinaryClassificationEvaluator.scala
<filename>kayenta-judge/src/main/scala/com/netflix/kayenta/judge/evaluation/BinaryClassificationEvaluator.scala /* * Copyright 2017 Netflix, Inc. * * Licensed under the Apache License, Version 2.0 (the "License") * you may not use this file except in compliance with the License. * You may obtain a copy of the Lice...
renhaocui/ensembleTopic
TMT/test2.scala
<reponame>renhaocui/ensembleTopic import scalanlp.io._; import scalanlp.stage._; import scalanlp.stage.text._; import scalanlp.text.tokenize._; import scalanlp.pipes.Pipes.global._; import edu.stanford.nlp.tmt.stage._; import edu.stanford.nlp.tmt.model.lda._; import edu.stanford.nlp.tmt.model.llda._; val modelPath = ...
renhaocui/ensembleTopic
TMT/train.scala
<gh_stars>1-10 import scalanlp.io._; import scalanlp.stage._; import scalanlp.stage.text._; import scalanlp.text.tokenize._; import scalanlp.pipes.Pipes.global._; import edu.stanford.nlp.tmt.stage._; import edu.stanford.nlp.tmt.model.lda._; import edu.stanford.nlp.tmt.model.llda._; val source = CSVFile("TMT\\LDAForma...
f-loris/scio-idea-plugin
src/main/scala/com/spotify/scio/ScioInjector.scala
/* * Copyright 2016 Spotify AB. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to ...
guizmaii/BooleanDsl
src/main/scala/com/guizmaii/boolean/dsl/BooleanDslV1.scala
<reponame>guizmaii/BooleanDsl package com.guizmaii.boolean.dsl sealed trait BooleanDslV1 { def unary_! : BooleanDslV1 final def &&(that: BooleanDslV1): BooleanDslV1 = BooleanDslV1.And(this, that) final def ||(that: BooleanDslV1): BooleanDslV1 = BooleanDslV1.Or(this, that) final def not: BooleanDslV1 ...
guizmaii/BooleanDsl
build.sbt
name := "BooleanDsl" version := "0.1" scalaVersion := "2.13.6" libraryDependencies ++= Seq( "org.scalatest" %% "scalatest" % "3.2.3" % Test, "org.scalatestplus" %% "scalacheck-1-14" % "3.2.2.0" % Test, "org.scalacheck" %% "scalacheck" % "1.15.1" % Test )
olka/stanoq
src/test/scala/org/stanoq/tests/crawler/ServiceNegativeSpec.scala
package org.stanoq.tests.crawler import akka.http.scaladsl.server.ValidationRejection import akka.http.scaladsl.testkit.ScalatestRouteTest import org.scalatest._ import org.stanoq.crawler.CrawlerService import org.stanoq.crawler.model.{ConfigProperties, CrawlerProtocols} class ServiceNegativeSpec extends FlatSpec wit...
olka/stanoq
src/test/scala/org/stanoq/load/tests/CrawlerLoadTest.scala
package org.stanoq.load.tests import io.gatling.core.Predef._ import io.gatling.http.Predef._ import scala.concurrent.duration._ class CrawlerLoadTest extends Simulation { val httpConf = http .baseURL("http://localhost:9000") .acceptHeader("application/json,application/xhtml+xml,application/xml;q=0.9,*/*...
olka/stanoq
src/test/scala/org/stanoq/tests/crawler/CrawlerSpec.scala
package org.stanoq.tests.crawler import org.scalatest._ import org.stanoq.crawler.Crawler import spray.json._ import org.stanoq.crawler.model.{ConfigProperties, Node} class CrawlerSpec extends FlatSpec with Matchers { "Crawler" should "handle recursive page structure" in { val crawler = new Crawler(ConfigPrope...
olka/stanoq
build.sbt
name := "stanoq" organization := "com.stanoq" version := "1.0" scalaVersion := "2.11.8" scalacOptions := Seq("-unchecked", "-deprecation", "-encoding", "utf8") enablePlugins(JavaAppPackaging) enablePlugins(GatlingPlugin) libraryDependencies ++= { val akkaV = "2.4.16" val akkaHttpV = "10.0.9" val scalaTe...
olka/stanoq
src/main/scala/org/stanoq/crawler/MongoHelper.scala
<gh_stars>1-10 package org.stanoq.crawler import java.util.concurrent.TimeUnit import com.typesafe.config.ConfigFactory import org.mongodb.scala.{MongoClient, MongoCollection} import org.stanoq.crawler.model._ import org.mongodb.scala.bson.codecs.Macros._ import org.mongodb.scala.bson.codecs.DEFAULT_CODEC_REGISTRY im...
olka/stanoq
src/test/scala/org/stanoq/load/tests/GatlingRunner.scala
<filename>src/test/scala/org/stanoq/load/tests/GatlingRunner.scala package org.stanoq.load.tests import io.gatling.app.Gatling import io.gatling.core.config.GatlingPropertiesBuilder object GatlingRunner extends App { val props = new GatlingPropertiesBuilder props.simulationClass(classOf[CrawlerLoadTest].getCanoni...
olka/stanoq
src/main/scala/org/stanoq/crawler/Crawler.scala
<reponame>olka/stanoq package org.stanoq.crawler import java.util.Collections import java.util.concurrent.{ConcurrentHashMap, TimeUnit} import akka.actor.ActorSystem import akka.event.Logging import org.jsoup.nodes.Document import org.jsoup.{Connection, HttpStatusException, Jsoup} import org.stanoq.crawler.model.{Con...
olka/stanoq
src/test/scala/org/stanoq/tests/crawler/ServiceSpec.scala
<filename>src/test/scala/org/stanoq/tests/crawler/ServiceSpec.scala package org.stanoq.tests.crawler import akka.event.NoLogging import akka.http.scaladsl.model.ContentTypes._ import akka.http.scaladsl.model.StatusCodes._ import akka.http.scaladsl.testkit.{RouteTestTimeout, ScalatestRouteTest} import org.scalatest._ i...
olka/stanoq
src/main/scala/org/stanoq/version/VersionService.scala
<reponame>olka/stanoq<gh_stars>1-10 package org.stanoq.version package org.stanoq.crawler import akka.http.scaladsl.model.{HttpEntity, _} import akka.http.scaladsl.server.Directives._ import com.typesafe.config.ConfigFactory class VersionService { val version = ConfigFactory.load().getString("stanoq.version") ...
olka/stanoq
project/plugins.sbt
addSbtPlugin("com.eed3si9n" % "sbt-assembly" % "0.14.3") addSbtPlugin("com.typesafe.sbt" % "sbt-native-packager" % "1.2.0-M7") addSbtPlugin("org.scoverage" % "sbt-scoverage" % "1.5.0") addSbtPlugin("io.gatling" % "gatling-sbt" % "2.2.0") addSbtPlugin("org.scoverage" % "sbt-coveralls" % "1.1.0")
olka/stanoq
src/test/scala/org/stanoq/load/tests/VersionLoadTest.scala
<reponame>olka/stanoq<filename>src/test/scala/org/stanoq/load/tests/VersionLoadTest.scala package org.stanoq.load.tests import io.gatling.core.Predef._ import io.gatling.http.Predef._ import scala.concurrent.duration._ class VersionLoadTest extends Simulation{ val httpConf = http.baseURL("http://stanoq.herokuapp....
olka/stanoq
src/main/scala/org/stanoq/CorsSupport.scala
package org.stanoq import akka.http.scaladsl.model.HttpMethods._ import akka.http.scaladsl.model.{StatusCodes, HttpResponse} import akka.http.scaladsl.model.headers._ import akka.http.scaladsl.server.Directives._ import akka.http.scaladsl.server.{Directive0, Route} trait CorsSupport { //this directive adds access ...
olka/stanoq
src/test/scala/org/stanoq/tests/VersionSpec.scala
<reponame>olka/stanoq<filename>src/test/scala/org/stanoq/tests/VersionSpec.scala package org.stanoq.tests import akka.http.scaladsl.model.ContentTypes._ import akka.http.scaladsl.model.StatusCodes._ import akka.http.scaladsl.testkit.ScalatestRouteTest import org.scalatest._ import org.stanoq.version.org.stanoq.crawler...
olka/stanoq
src/main/scala/org/stanoq/RestController.scala
package org.stanoq import akka.actor.ActorSystem import akka.event.Logging import akka.http.scaladsl.Http import akka.http.scaladsl.server.Directives._ import akka.http.scaladsl.server.directives.DebuggingDirectives import akka.stream.ActorMaterializer import com.typesafe.config.ConfigFactory import org.stanoq.crawler...
olka/stanoq
src/main/scala/org/stanoq/auth/JwtAuth.scala
<filename>src/main/scala/org/stanoq/auth/JwtAuth.scala package org.stanoq.auth import pdi.jwt.{Jwt, JwtAlgorithm} /** * Jwt helper class * @param payload @Json object converted to @String * @param secret Combination of url and depth. This is UUID of crawling result */ case class JwtAuth(payload:String, secret...
olka/stanoq
src/main/scala/org/stanoq/crawler/StreamService.scala
<reponame>olka/stanoq package org.stanoq.crawler import akka.actor.ActorSystem import akka.http.scaladsl.common.{EntityStreamingSupport, JsonEntityStreamingSupport} import akka.http.scaladsl.server.Directives._ import akka.stream.ThrottleMode import akka.stream.impl.Stages.DefaultAttributes import akka.stream.scaladsl...
olka/stanoq
src/main/scala/org/stanoq/crawler/CrawlerService.scala
package org.stanoq.crawler import akka.actor.ActorSystem import akka.http.scaladsl.model._ import akka.http.scaladsl.server.Directives._ import org.stanoq.crawler.model._ import spray.json._ import scala.concurrent._ import scala.concurrent.Future class CrawlerService() extends CrawlerProtocols { implicit val blo...
olka/stanoq
src/main/scala/org/stanoq/crawler/model/ConfigProperties.scala
<gh_stars>1-10 package org.stanoq.crawler.model import java.awt.Color import java.net.{URI, URL} import akka.http.scaladsl.marshallers.sprayjson.SprayJsonSupport import spray.json.{DefaultJsonProtocol, RootJsonFormat} import scala.collection.mutable.Set import scala.util.Try case class ConfigProperties(url: String,...
olka/stanoq
src/test/scala/org/stanoq/tests/crawler/AuthSpec.scala
package org.stanoq.tests.crawler import org.scalatest._ import org.stanoq.auth.JwtAuth import spray.json._ import org.stanoq.crawler.model.{ConfigProperties, CrawlerProtocols} import pdi.jwt.exceptions.JwtValidationException class AuthSpec extends FlatSpec with Matchers with CrawlerProtocols { val config = ConfigP...
Frugghi/TweetSpark
tweet-spark-scala/src/test/scala/com/tommasomadonia/spark/test/WordCountTests.scala
package com.tommasomadonia.spark.test import com.tommasomadonia.spark.Words import org.scalatest.{Matchers, GivenWhenThen, FlatSpec} class WordCountTests extends FlatSpec with JSONSchemaSpec with GivenWhenThen with Matchers { "Empty JSON" should "have no words" in { Given("an empty DataFrame") val dataFram...
Frugghi/TweetSpark
tweet-spark-scala/src/main/scala/com/tommasomadonia/spark/TwitterAnalyzer.scala
package com.tommasomadonia.spark import org.apache.hadoop.conf.Configuration import org.apache.hadoop.fs.{FileSystem, Path} import org.apache.spark.sql.SQLContext import org.apache.spark.{SparkContext, SparkConf} object TwitterAnalyzer { def main(args: Array[String]) { if (args.length < 1) { System.err.p...
Frugghi/TweetSpark
statsd-jvm-profiler/example/StatsDProfilerFlowListener.scala
package com.etsy.cascading.flow import java.util.Properties import cascading.flow.{Flow, FlowListener, FlowStep} import org.apache.hadoop.mapred.JobConf import scala.collection.JavaConversions._ /** * Flow listener for setting up JobConf to enable statsd-jvm-profiler */ class StatsDProfilerFlowListener extends Fl...
Frugghi/TweetSpark
tweet-spark-scala/src/test/scala/com/tommasomadonia/spark/test/SparkSpec.scala
package com.tommasomadonia.spark.test import org.apache.log4j.{Level, Logger} import org.apache.spark._ import org.scalatest._ trait SparkSpec extends BeforeAndAfterAll { this: Suite => private val master = "local[2]" private val appName = this.getClass.getSimpleName private var _sparkContext: SparkContext ...
Frugghi/TweetSpark
tweet-spark-scala/src/main/scala/com/tommasomadonia/spark/ActiveTweeters.scala
package com.tommasomadonia.spark import org.apache.spark.sql.{DataFrame, SQLContext} object ActiveTweeters { def find(sqlContext: SQLContext, table: String): DataFrame = { sqlContext.sql(s""" |SELECT user.screen_name, COUNT(*) AS total_count |FROM $table ...
Frugghi/TweetSpark
tweet-spark-scala/src/main/scala/com/tommasomadonia/spark/TweetDataFrame.scala
<filename>tweet-spark-scala/src/main/scala/com/tommasomadonia/spark/TweetDataFrame.scala package com.tommasomadonia.spark import org.apache.spark.sql.{Row, DataFrame} import org.apache.spark.sql.functions._ import scala.collection.mutable.ArrayBuffer private case class _Indices(indices: Array[Long]) package object ...
Frugghi/TweetSpark
tweet-spark-scala/src/test/scala/com/tommasomadonia/spark/test/SparkSQLSpec.scala
<reponame>Frugghi/TweetSpark package com.tommasomadonia.spark.test import org.apache.spark.sql.SQLContext import org.scalatest.Suite trait SparkSQLSpec extends SparkSpec { this: Suite => private var _sqlContext: SQLContext = _ def sqlContext = _sqlContext override def beforeAll(): Unit = { super.before...
Frugghi/TweetSpark
tweet-spark-scala/src/test/scala/com/tommasomadonia/spark/test/JSONSchemaSpec.scala
package com.tommasomadonia.spark.test import org.apache.spark.sql.DataFrameReader import org.apache.spark.sql.types.StructType import org.scalatest.Suite trait JSONSchemaSpec extends SparkSQLSpec { this: Suite => private val jsonSchema = "test/schema.json" private var _schema: StructType = _ private var _da...
Frugghi/TweetSpark
tweet-spark-scala/src/main/scala/com/tommasomadonia/spark/Tweet.scala
package com.tommasomadonia.spark import scala.collection.mutable.ArrayBuffer case class Tweet(text: String, indices: Array[(Long, Long)]) { def tokenize(): TraversableOnce[String] = this match { case Tweet(tweet, indices) if (tweet == null || tweet.isEmpty) => Array[String]() case Tweet(tweet, indices) => ...
Frugghi/TweetSpark
tweet-spark-scala/src/main/scala/com/tommasomadonia/spark/Words.scala
package com.tommasomadonia.spark import org.apache.spark.rdd.RDD import org.apache.spark.sql.{DataFrame, Row} import org.apache.spark.sql.functions._ import scala.collection.mutable.{ArrayBuffer, StringBuilder} import com.github.nscala_time.time.Imports._ import com.tommasomadonia.spark.dataframe_extension._ priva...
skygoo/octo
src/main/scala/org/seekloud/octo/ptcl/IceProtocol.scala
package org.seekloud.octo.ptcl /** * Created by sky * Date on 2019/8/16 * Time at 17:48 */ object IceProtocol { case class CandidateInfo( candidate: String, sdpMid: String, sdpMLineIndex: Int ) }
skygoo/octo
src/main/scala/org/seekloud/octo/http/SessionBase.scala
package org.seekloud.octo.http import akka.http.scaladsl.model._ import akka.http.scaladsl.server import akka.http.scaladsl.server.Directives.{complete, extractRequestContext, onComplete, redirect, reject} import akka.http.scaladsl.server._ import akka.http.scaladsl.server.directives.BasicDirectives import org.seeklou...
skygoo/octo
src/main/scala/org/seekloud/octo/http/HttpService.scala
package org.seekloud.octo.http import akka.actor.{ActorSystem, Scheduler} import akka.http.scaladsl.server.Directives._ import akka.http.scaladsl.server.Route import akka.stream.Materializer import akka.util.Timeout import akka.actor.typed.scaladsl.AskPattern._ import scala.concurrent.Future import scala.concurrent.E...
skygoo/octo
src/main/scala/org/seekloud/octo/common/AppSettings.scala
// Copyright 2018 seekloud (https://github.com/seekloud) // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required b...
skygoo/octo
src/main/scala/org/seekloud/octo/bridge/DtlsHandler.scala
<gh_stars>0 package org.seekloud.octo.bridge import java.io.IOException import javax.sdp.{MediaDescription, SdpException, SdpParseException, SessionDescription} import org.ice4j.ice.{CandidatePair, Component, IceMediaStream, IceProcessingState} import org.seekloud.octo.bridge.dtls.mock.Connect import org.seekloud.oct...
skygoo/octo
build.sbt
<filename>build.sbt name := "octo" version := "0.1" scalaVersion := "2.12.8" //resolvers += Resolver.mavenLocal resolvers += "Jisti ORG Snapshots" at "https://github.com/jitsi/jitsi-maven-repository/raw/master/snapshots/" //resolvers += "Jisti ORG Releases" at "https://github.com/jitsi/jitsi-maven-repository/raw/m...
skygoo/octo
src/main/scala/org/seekloud/octo/ptcl/BrowserMsg.scala
<reponame>skygoo/octo package org.seekloud.octo.ptcl import org.seekloud.octo.ptcl.IceProtocol.CandidateInfo /** * Created by sky * Date on 2019/8/16 * Time at 16:56 */ object BrowserMsg { object MsgId { val PING = "PING" //前端定时发送 val PONG = "PONG" //后台回复 val Anchor_SDP_OFFER = "Anchor_SDP_OFFE...
skygoo/octo
src/main/scala/org/seekloud/octo/bridge/IceHandler.scala
<filename>src/main/scala/org/seekloud/octo/bridge/IceHandler.scala<gh_stars>0 package org.seekloud.octo.bridge import java.beans.{PropertyChangeEvent, PropertyChangeListener} import java.io.IOException import javax.sdp.{MediaDescription, SdpException, SessionDescription} import org.ice4j.{Transport, TransportAddress}...
skygoo/octo
src/main/scala/org/seekloud/octo/ptcl/EpInfo.scala
package org.seekloud.octo.ptcl /** * Created by sky * Date on 2019/8/18 * Time at 14:24 * distinguish endpoint webSocketSession with info */ case class EpInfo(id:String)
skygoo/octo
src/main/scala/org/seekloud/octo/core/EndPointManager.scala
<reponame>skygoo/octo<filename>src/main/scala/org/seekloud/octo/core/EndPointManager.scala package org.seekloud.octo.core import akka.actor.typed.{ActorRef, Behavior} import akka.actor.typed.scaladsl.{ActorContext, Behaviors, TimerScheduler} import akka.http.scaladsl.model.ws.{Message, TextMessage} import akka.stream....
skygoo/octo
project/Dependencies.scala
<filename>project/Dependencies.scala import sbt._ /** * User: Taoz * Date: 6/13/2017 * Time: 9:38 PM */ object Dependencies { val slickV = "3.2.3" val akkaV = "2.5.22" val akkaHttpV = "10.1.8" val scalaXmlV = "1.1.0" val circeVersion = "0.9.3" val scalaJsDomV = "0.9.6" val akkaSeq = Seq( "...
skygoo/octo
src/main/scala/org/seekloud/octo/core/EndPointWorker.scala
<reponame>skygoo/octo<gh_stars>0 package org.seekloud.octo.core import java.io.FileInputStream import akka.actor.typed.{ActorRef, Behavior} import akka.actor.typed.scaladsl.{Behaviors, StashBuffer, TimerScheduler} import akka.stream.OverflowStrategy import akka.stream.scaladsl.Flow import akka.stream.typed.scaladsl.{...
skygoo/octo
src/main/scala/org/seekloud/octo/http/SocketService.scala
<gh_stars>0 package org.seekloud.octo.http import akka.http.scaladsl.model.ws.Message import akka.http.scaladsl.server.Directives._ import org.slf4j.LoggerFactory import akka.actor.typed.scaladsl.AskPattern._ import akka.http.scaladsl.server.Route import akka.stream.scaladsl.Flow import scala.concurrent.Future import...
skygoo/octo
src/test/scala/org/seekloud/octo/Test.scala
package org.seekloud.octo import java.beans.PropertyChangeEvent import java.beans.PropertyChangeListener import java.io.BufferedReader import java.io.InputStreamReader import java.net.DatagramSocket import java.net.SocketAddress import java.util import org.ice4j.Transport import org.ice4j.TransportAddress import org.i...
skygoo/octo
src/main/scala/org/seekloud/octo/Boot.scala
<filename>src/main/scala/org/seekloud/octo/Boot.scala package org.seekloud.octo import java.io.FileInputStream import java.security.{KeyStore, SecureRandom} import akka.actor.ActorSystem import akka.actor.typed.ActorRef import akka.actor.typed.scaladsl.adapter._ import akka.dispatch.MessageDispatcher import akka.even...
skygoo/octo
src/main/scala/org/seekloud/octo/ptcl/WebSocketSession.scala
package org.seekloud.octo.ptcl import akka.actor.typed.ActorRef import org.seekloud.octo.core.EndPointWorker /** * Created by sky * Date on 2019/8/16 * Time at 15:16 */ case class WebSocketSession( id: String, actor:ActorRef[EndPointWorker.Command], ...
Axiometry/agario-client
src/main/scala/me/axiometry/agario/Agario.scala
<gh_stars>1-10 package me.axiometry.agario import akka.io.IO import akka.pattern.ask import akka.actor.ActorSystem import akka.util.Timeout import spray.can.Http import spray.http._ import HttpHeaders._ import HttpMethods._ import MediaTypes._ import spray.client.pipelining._ import scala.concurrent._ import scala.c...
Axiometry/agario-client
src/main/scala/me/axiometry/agario/Entity.scala
<filename>src/main/scala/me/axiometry/agario/Entity.scala package me.axiometry.agario trait Entity { def world: World def x: Double def x_=(x: Double): Unit def y: Double def y_=(y: Double): Unit }
Axiometry/agario-client
src/main/scala/me/axiometry/agario/net/DataView.scala
<reponame>Axiometry/agario-client package me.axiometry.agario.net import akka.util.ByteString case class DataView(array: Array[Byte], offset: Int, length: Int)(implicit val byteOrder: ByteOrder = ByteOrder.BigEndian) { import ByteOrder._ def getInt8(off: Int)(implicit byteOrder: ByteOrder = byteOrder): Byte = ...
Axiometry/agario-client
src/main/scala/me/axiometry/agario/Client.scala
<filename>src/main/scala/me/axiometry/agario/Client.scala package me.axiometry.agario import me.axiometry.agario.net._ import akka.actor.{ ActorSystem, Props } import grizzled.slf4j.Logging import spray.can.Http import spray.can.websocket.frame._ import spray.http._ import spray.httpx.RequestBuilding._ case class Cl...
Axiometry/agario-client
src/main/scala/me/axiometry/agario/Server.scala
<filename>src/main/scala/me/axiometry/agario/Server.scala package me.axiometry.agario import akka.actor.ActorSystem import scala.util._ case class ServerAddress(host: String, port: Int) object ServerAddress { private[this] val ServerPattern = "([^:]+):([0-9]+)".r def unapply(string: String): Option[ServerAddress...
Axiometry/agario-client
src/main/scala/me/axiometry/agario/World.scala
<reponame>Axiometry/agario-client<filename>src/main/scala/me/axiometry/agario/World.scala package me.axiometry.agario class World { }
Axiometry/agario-client
src/main/scala/me/axiometry/agario/net/WebSocketClient.scala
package me.axiometry.agario.net import akka.actor.ActorSystem import akka.io.IO import spray.can.Http import spray.can.server.UHttp import spray.can.websocket._ import spray.can.websocket.frame._ import spray.http.HttpRequest abstract class WebSocketClient(connect: Http.Connect, val upgradeRequest: HttpRequest)(impli...
Axiometry/agario-client
src/main/scala/me/axiometry/agario/Carb.scala
<reponame>Axiometry/agario-client package me.axiometry.agario case class Carb(override val world: World) extends Entity { override var x: Double = _ override var y: Double = _ }
sequencer/rocket-dsp-utils
src/main/scala/freechips/rocketchip/jtag2mm/JtagFuzzer.scala
// SPDX-License-Identifier: Apache-2.0 package freechips.rocketchip.jtag2mm import chisel3._ import chisel3.experimental._ import chisel3.util._ import chisel3.util.random.LFSR class InvertedJtagIO extends Bundle { // TRST (4.6) is optional and not currently implemented. val TCK = Output(Bool()) val TMS = Out...
sequencer/rocket-dsp-utils
build.sbt
<gh_stars>1-10 // SPDX-License-Identifier: Apache-2.0 git.remoteRepo := "<EMAIL>:ucb-bar/dsptools.git" enablePlugins(SiteScaladocPlugin) enablePlugins(GhpagesPlugin) def scalacOptionsVersion(scalaVersion: String): Seq[String] = { Seq() ++ { // If we're building with Scala > 2.11, enable the compile option ...
sequencer/rocket-dsp-utils
src/test/scala/freechips/rocketchip/jtag2mm/JtagFuzzerTester.scala
<filename>src/test/scala/freechips/rocketchip/jtag2mm/JtagFuzzerTester.scala // SPDX-License-Identifier: Apache-2.0 package freechips.rocketchip.jtag2mm import dsptools.DspTester import org.scalatest.flatspec.AnyFlatSpec import org.scalatest.matchers.should.Matchers class JtagFuzzerTester(dut: JtagFuzzer) extends Ds...
sequencer/rocket-dsp-utils
src/main/scala/package.scala
<filename>src/main/scala/package.scala //val config = chipsalliance.rocketchip.config
sequencer/rocket-dsp-utils
src/main/scala/freechips/rocketchip/amba/axi4stream/SimpleSplitter.scala
package freechips.rocketchip.amba.axi4stream import chisel3._ import freechips.rocketchip.amba.axi4stream.{AXI4StreamMasterPortParameters, AXI4StreamNexusNode, AXI4StreamSlavePortParameters} import freechips.rocketchip.config.Parameters import freechips.rocketchip.diplomacy._ class SimpleSplitter() extends LazyModule...
sequencer/rocket-dsp-utils
src/main/scala/freechips/rocketchip/interrupts/Nodes.scala
package freechips.rocketchip.interrupts import chisel3._ import chipsalliance.rocketchip.config.Parameters import freechips.rocketchip.diplomacy._ object IntBundleBridgeImp extends BundleBridgeImp[Vec[Bool]] case class IntToBundleBridgeNode(sinkParams: IntSinkPortParameters)(implicit valName: ValName) extends MixedA...
sequencer/rocket-dsp-utils
src/main/scala/freechips/rocketchip/jtag2mm/Utils.scala
// See ./LICENSE for license details. package freechips.rocketchip.jtag2mm import chisel3._ import chisel3.util._ // This code was taken from https://github.com/ucb-art/chisel-jtag/blob/master/src/main/scala/jtag/Utils.scala and adjusted to our design needs /** Bundle representing a tristate pin. */ class Tristat...
sequencer/rocket-dsp-utils
src/test/scala/dspblocks/DspBlockFirSpec.scala
<reponame>sequencer/rocket-dsp-utils<filename>src/test/scala/dspblocks/DspBlockFirSpec.scala // SPDX-License-Identifier: Apache-2.0 package dspblocks import chisel3._ import chisel3.iotesters._ import chisel3.util.Cat import freechips.rocketchip.amba.axi4._ import freechips.rocketchip.amba.axi4stream._ import freechi...
sequencer/rocket-dsp-utils
src/test/scala/dsptools/tester/MemMasterSpec.scala
package dsptools.tester import chisel3._ import chisel3.iotesters.PeekPokeTester import freechips.rocketchip.amba.apb._ import freechips.rocketchip.amba.axi4._ import freechips.rocketchip.config.Parameters import freechips.rocketchip.diplomacy._ import freechips.rocketchip.interrupts._ import freechips.rocketchip.regm...
tumsgis/veganmentor
src/test/scala/VeganMentorDbIntegrationTest.scala
<reponame>tumsgis/veganmentor import java.time.LocalDateTime import DataStructure.{Mentee, Mentor} import junit.framework.TestCase import org.junit.Assert._ import org.junit.{After, Before, Test} class VeganMentorDbIntegrationTest extends TestCase { @Before override def setUp(): Unit = MongoDbRepo.dropAllPartic...
tumsgis/veganmentor
src/main/scala/VeganMentor.scala
import DataStructure._ import scala.annotation.tailrec object VeganMentor { def saveInputFileToDb(csvFilePath: String): Unit = { val (mentors, mentees) = Parser.parseInputFile(csvFilePath) mentors.foreach(MongoDbRepo.saveParticipant) mentees.foreach(MongoDbRepo.saveParticipant) } def process: Unit...
tumsgis/veganmentor
src/test/scala/VeganMentorTest.scala
import junit.framework.TestCase import org.junit.Assert._ import org.junit.{Before, Test} import MongoDbRepo._ class VeganMentorTest extends TestCase{ @Before override def setUp(): Unit = dropAllParticipants() @Test def testProcessing01(): Unit = { val testFile = "src/test/files/Vegan_mentor_test_01.csv"...
tumsgis/veganmentor
src/main/scala/Parser.scala
import java.io.File import DataStructure.{Mentee, Mentor, SortedQueue, Participant} import com.github.tototoshi.csv.CSVReader object Parser { def parseInputFile(filePath: String): (Seq[Mentor], Seq[Mentee]) = { val reader = CSVReader.open(new File (filePath)) val participants: Seq[Participant] = reader.al...
tumsgis/veganmentor
src/main/scala/Util.scala
import java.time.{LocalDateTime, ZoneId} import java.time.format.DateTimeFormatter import java.util.Date object Util { private val formatter = DateTimeFormatter.ofPattern("yyyy/MM/dd hh:mm:ss a z") private val formatterWithSimplerHourFormat = DateTimeFormatter.ofPattern("yyyy/MM/dd h:mm:ss a z") private def ge...
tumsgis/veganmentor
src/main/scala/DataStructure.scala
import java.time.LocalDateTime import com.mongodb.casbah.Imports object DataStructure { val mentorShipMap = Map("Mentor" -> true, "Lærlingur" -> false) sealed trait Participant { def timestamp: LocalDateTime def email: String def name: String def note: String def approvedTermsAndConditions: ...
tumsgis/veganmentor
build.sbt
name := "VeganMentor" version := "0.1" scalaVersion := "2.12.4" libraryDependencies += "com.github.tototoshi" %% "scala-csv" % "1.3.5" libraryDependencies += "junit" % "junit" % "4.12" % Test libraryDependencies += "com.google.apis" % "google-api-services-gmail" % "v1-rev75-1.23.0" libraryDependencies += "com.goog...