repo_name stringlengths 6 97 | path stringlengths 3 341 | text stringlengths 8 1.02M |
|---|---|---|
parwen68/advent-of-code-2017 | shared/src/test/scala/se/parwen/adoc/day10/SolverSpec.scala | package se.parwen.adoc.day10
import org.scalatest.{FlatSpec, Matchers}
class SolverSpec extends FlatSpec with Matchers {
"The solver" should "handle 1" in {
import Solver._
val r = swap(List(0, 1, 2, 3, 4), 0, 3)
r should be (List(2, 1, 0, 3, 4))
}
"The solver" should "handle 2" in {
import S... |
parwen68/advent-of-code-2017 | shared/src/main/scala/se/parwen/adoc/day10/Solver.scala | <reponame>parwen68/advent-of-code-2017
package se.parwen.adoc.day10
import se.parwen.adoc.day10.Solver.solveForList
import scala.annotation.tailrec
object Solver {
def swap(l: List[Int], from: Int, length: Int) : List[Int] = {
val nFrom = from % l.length
val nTo = (from + length) % l.length
if (lengt... |
parwen68/advent-of-code-2017 | shared/src/main/scala/se/parwen/adoc/day8/Solver.scala | <reponame>parwen68/advent-of-code-2017
package se.parwen.adoc.day8
object Solver {
val pattern = raw"(\w+) (\w+) ([-\d]+) if (\w+ [!><=]+ [-\d]+)".r
val patternCond = raw"(\w+) ([!><=]+) ([-\d]+)".r
case class Exp(reg: String, cmd: String, amount: Int, cond: String)
def parse(input: String) = {
val row... |
parwen68/advent-of-code-2017 | shared/src/main/scala/se/parwen/adoc/day14/Solver.scala | package se.parwen.adoc.day14
import scala.annotation.tailrec
object Solver {
def toAscii(s: String): List[Int] = {
s.map(_.toInt).toList
}
def swap(l: List[Int], from: Int, length: Int) : List[Int] = {
val nFrom = from % l.length
val nTo = (from + length) % l.length
if (length == 0) {
l... |
parwen68/advent-of-code-2017 | shared/src/test/scala/se/parwen/adoc/day19/SolverSpec.scala | package se.parwen.adoc.day19
import org.scalatest._
class SolverSpec extends FlatSpec with Matchers {
"Solver parse" should "return 14x6 array" in {
val a = Solver.parse(""" |
$ | +--+
$ A | C
$ F---|----E|--+
... |
parwen68/advent-of-code-2017 | shared/src/test/scala/se/parwen/adoc/day5/SolverSpec.scala | package se.parwen.adoc.day5
import org.scalatest._
class SolverSpec extends FlatSpec with Matchers {
import Solver._
"The solver step1" should "return 5 for 0 3 0 1 -3" in {
solveStep1(Array(0,3,0,1,-3)) should be (5)
}
"The solver step2" should "return 10 for 0 3 0 1 -3" in {
solveStep2(Ar... |
parwen68/advent-of-code-2017 | shared/src/main/scala/se/parwen/adoc/day14/Input.scala | <reponame>parwen68/advent-of-code-2017
package se.parwen.adoc.day14
object Input {
val input = "nbysizxe"
}
|
parwen68/advent-of-code-2017 | shared/src/main/scala/se/parwen/adoc/day12/Solver.scala | package se.parwen.adoc.day12
import scala.annotation.tailrec
object Solver {
case class Program(name: String, programs: List[Program])
private def parse(input: String) = {
val rows = input.split('\n').map { v =>
val e = v.split("<->")
(e(0).trim.toInt, e(1).split(',').map(s => s.trim.toInt).toLi... |
parwen68/advent-of-code-2017 | shared/src/test/scala/se/parwen/adoc/day12/SolverSpec.scala | package se.parwen.adoc.day12
import org.scalatest.{FlatSpec, Matchers}
class SolverSpec extends FlatSpec with Matchers {
"The solver step 1" should "return 6" in {
import Solver._
val r = solveStep1("""0 <-> 2
|1 <-> 1
|2 <-> 0, 3, 4
... |
parwen68/advent-of-code-2017 | shared/src/main/scala/se/parwen/adoc/day11/Solver.scala | <reponame>parwen68/advent-of-code-2017<filename>shared/src/main/scala/se/parwen/adoc/day11/Solver.scala
package se.parwen.adoc.day11
/*
s + ne = se
s + nw = sw
n + se = ne
n + sw = nw
---
se + sw = s
ne + nw = n
---
s + n = 0
sw + ne = 0
se + nw = 0
---
s,sw,se = s, s
s,sw,nw = sw, sw
s,ne,se = se, s... |
parwen68/advent-of-code-2017 | shared/src/test/scala/se/parwen/adoc/day8/SolverSpec.scala | package se.parwen.adoc.day8
import org.scalatest.{FlatSpec, Matchers}
class SolverSpec extends FlatSpec with Matchers {
"The solver" should "return a and 1" in {
import Solver._
val (reg, v) = solveStep1("""b inc 5 if a > 1
|a inc 1 if b < 5
|c dec -10 if a >= 1
... |
parwen68/advent-of-code-2017 | shared/src/main/scala/se/parwen/adoc/day5/Solver.scala | package se.parwen.adoc.day5
object Solver {
def solveStep1(list: Array[Int]): Int = {
var ptr = 0
var prevPtr = ptr
import scala.util.control.Breaks._
var cnt = 0
breakable {
while (true) {
cnt += 1
val v = list(ptr)
ptr += v
list(prevPtr) = list(prevPtr) ... |
parwen68/advent-of-code-2017 | shared/src/test/scala/se/parwen/adoc/day15/SolverSpec.scala | <reponame>parwen68/advent-of-code-2017<filename>shared/src/test/scala/se/parwen/adoc/day15/SolverSpec.scala<gh_stars>0
package se.parwen.adoc.day15
import org.scalatest.{FlatSpec, Matchers}
class SolverSpec extends FlatSpec with Matchers {
"Solver step1" should "return 588" in {
Solver.solveStep1(65, 8921) sho... |
parwen68/advent-of-code-2017 | shared/src/main/scala/se/parwen/adoc/day10/Input.scala | <reponame>parwen68/advent-of-code-2017
package se.parwen.adoc.day10
object Input {
val input = List(18,1,0,161,255,137,254,252,14,95,165,33,181,168,2,188)
val inputAsStr = "18,1,0,161,255,137,254,252,14,95,165,33,181,168,2,188"
}
|
parwen68/advent-of-code-2017 | shared/src/main/scala/se/parwen/adoc/day7/Solver.scala | <filename>shared/src/main/scala/se/parwen/adoc/day7/Solver.scala
package se.parwen.adoc.day7
import se.parwen.adoc.day7.Solver.{solveStep1, solveStep2}
import scala.annotation.tailrec
object Solver {
val pattern = raw"(\w+) \((\d+)\)[-> ]*([\w, ]*)".r
case class Program(name: String, weight: Int = -1, ontopof:... |
parwen68/advent-of-code-2017 | shared/src/main/scala/se/parwen/adoc/day16/Solver.scala | <gh_stars>0
package se.parwen.adoc.day16
import scala.annotation.tailrec
object Solver {
def spin(inp: Array[Char], steps: Int) : Array[Char] = {
inp.takeRight(steps) ++ inp.dropRight(steps)
}
def exchange(inp: Array[Char], pos1: Int, pos2: Int) : Array[Char] = {
inp.updated(pos1, inp(pos2)).updated(p... |
parwen68/advent-of-code-2017 | shared/src/main/scala/se/parwen/adoc/day9/Solver.scala | <filename>shared/src/main/scala/se/parwen/adoc/day9/Solver.scala<gh_stars>0
package se.parwen.adoc.day9
object Solver {
def solveStep1(input: String): Int = {
case class Context(ingarbage: Boolean = false, score: Int = 0, started: Int = 0){
def startGroup() = this.copy(started = started + 1)
def end... |
parwen68/advent-of-code-2017 | shared/src/test/scala/se/parwen/adoc/day16/SolverSpec.scala | <filename>shared/src/test/scala/se/parwen/adoc/day16/SolverSpec.scala
package se.parwen.adoc.day16
import org.scalatest._
import se.parwen.adoc.day16.Solver.{Exchange, Partner, Spin}
class SolverSpec extends FlatSpec with Matchers {
"Solver spin 3" should "return cdeab for abcde" in {
Solver.spin("abcde".toArr... |
parwen68/advent-of-code-2017 | shared/src/main/scala/se/parwen/adoc/day17/Solver.scala | package se.parwen.adoc.day17
object Solver {
def move(currPos: Int, steps: Int, bufferSize: Int): Int = {
val laps = steps / bufferSize
val reminder = steps % bufferSize
if(currPos + reminder >= bufferSize) {
currPos + reminder - bufferSize
} else {
currPos + reminder
}
}
def i... |
parwen68/advent-of-code-2017 | shared/src/test/scala/se/parwen/adoc/day14/SolverSpec.scala | package se.parwen.adoc.day14
import org.scalatest.{FlatSpec, Matchers}
class SolverSpec extends FlatSpec with Matchers {
/*
##.#.#.. d4
.#.#.#.# 55
....#.#. 0c
#.#.##.#
.##.#...
##..#..#
.#...#..
##.#.##.
*/
"a" should "b" in {
val a = Solver.solveStep1("flqrgnkx")
}
}
|
parwen68/advent-of-code-2017 | shared/src/main/scala/se/parwen/adoc/day8/Input.scala | <reponame>parwen68/advent-of-code-2017
package se.parwen.adoc.day8
object Input {
val input = """js inc 257 if wn < 9
|jq dec -586 if esb != -3
|gcf inc -603 if i >= -9
|gcf dec -300 if d != 1
|g inc -973 if gy > -1
|epp dec -79 if rjx ... |
parwen68/advent-of-code-2017 | shared/src/test/scala/se/parwen/adoc/day13/SolverSpec.scala | <reponame>parwen68/advent-of-code-2017
package se.parwen.adoc.day13
import org.scalatest.{FlatSpec, Matchers}
import Solver._
class SolverSpec extends FlatSpec with Matchers {
"The layer" should "return 0,1,2,3,2,1,0,1,2,3 in 0,1,2,3,4,5,6,7,8,9" in {
val layer = Layer.parse("0 : 4")
List(0,1,2,3,4,5,6,7,8... |
parwen68/advent-of-code-2017 | shared/src/main/scala/se/parwen/adoc/day4/Input.scala | <filename>shared/src/main/scala/se/parwen/adoc/day4/Input.scala
package se.parwen.adoc.day4
object Input {
val input =
"""
|kvvfl kvvfl olud wjqsqa olud frc
|slhm rdfm yxb rsobyt rdfm
|pib wzfr xyoakcu zoapeze rtdxt rikc jyeps wdyo hawr xyoakcu hawr
|ismtq qwoi kzt ktgzoc gnxblp dzfayil f... |
parwen68/advent-of-code-2017 | jvm/src/main/scala/se/parwen/adoc/AdocJVM.scala | <filename>jvm/src/main/scala/se/parwen/adoc/AdocJVM.scala
package se.parwen.adoc
import se.parwen.adoc.day10.Solver10
import se.parwen.adoc.day11.Solver11
import se.parwen.adoc.day12.Solver12
import se.parwen.adoc.day13.Solver13
import se.parwen.adoc.day14.Solver14
import se.parwen.adoc.day15.Solver15
import se.parwen... |
parwen68/advent-of-code-2017 | shared/src/test/scala/se/parwen/adoc/day11/SolverSpec.scala | package se.parwen.adoc.day11
import org.scalatest.{FlatSpec, Matchers}
class SolverSpec extends FlatSpec with Matchers {
"The solver step 1" should "return 3 for ne,ne,ne" in {
import Solver._
val r = solveStep1("ne,ne,ne")
r should be (3)
}
"The solver step 1" should "return 0 for ne,ne,sw,sw" in... |
parwen68/advent-of-code-2017 | shared/src/test/scala/se/parwen/adoc/day4/SolverSpec.scala | <filename>shared/src/test/scala/se/parwen/adoc/day4/SolverSpec.scala
package se.parwen.adoc.day4
import org.scalatest._
class SolverSpec extends FlatSpec with Matchers {
import Solver._
"The solver step 1" should "return valid for aa bb cc dd ee" in {
solveStep1("aa bb cc dd ee") should be(1)
}
"The so... |
parwen68/advent-of-code-2017 | shared/src/main/scala/se/parwen/adoc/day15/Input.scala | package se.parwen.adoc.day15
object Input {
val input = Map('A' -> 722, 'B' -> 354)
}
|
parwen68/advent-of-code-2017 | shared/src/main/scala/se/parwen/adoc/day13/Solver.scala | package se.parwen.adoc.day13
import scala.annotation.tailrec
object Solver {
case class Layer(pos: Int, depth: Int) {
val stream: List[Int] = (0 until depth-1).toList ++ (depth-1 until 0 by -1).toList
def getForPicoSec(pico: Int): Int = {
stream(pico % stream.length)
}
}
object Layer {
d... |
parwen68/advent-of-code-2017 | shared/src/main/scala/se/parwen/adoc/day20/Input.scala | <reponame>parwen68/advent-of-code-2017
package se.parwen.adoc.day20
object Input {
val input1 = """p=<1199,-2918,1457>, v=<-13,115,-8>, a=<-7,8,-10>
|p=<2551,2418,-1471>, v=<-106,-108,39>, a=<-6,-5,6>
|p=<-73,1626,1321>, v=<58,-118,-8>, a=<-6,2,-9>
|p=<-3297,-894,-551>... |
parwen68/advent-of-code-2017 | shared/src/test/scala/se/parwen/adoc/day17/SolverSpec.scala | <reponame>parwen68/advent-of-code-2017
package se.parwen.adoc.day17
import org.scalatest._
class SolverSpec extends FlatSpec with Matchers {
"The Solver move 3 from pos 0 in buffer [0]" should "be 0" in {
val buffer = List(0)
Solver.move(0, 3, buffer.size) should be (0)
}
"The Solver move 3 from pos 1... |
parwen68/advent-of-code-2017 | shared/src/main/scala/se/parwen/adoc/day19/Solver.scala | package se.parwen.adoc.day19
object Solver {
sealed trait Direction
trait HDirection extends Direction
trait VDirection extends Direction
case object Up extends VDirection
case object Down extends VDirection
case object Left extends HDirection
case object Right extends HDirection
case class Path(arr:... |
parwen68/advent-of-code-2017 | shared/src/main/scala/se/parwen/adoc/day6/Solver.scala | <filename>shared/src/main/scala/se/parwen/adoc/day6/Solver.scala
package se.parwen.adoc.day6
object Solver {
def stream(value: Int, length: Int, count: Int = 0) : Stream[Int] = {
if (count == length) Stream.empty
else Math.ceil(value.asInstanceOf[Double] / length).asInstanceOf[Int] #:: stream(value-1, lengt... |
parwen68/advent-of-code-2017 | shared/src/main/scala/se/parwen/adoc/day17/Input.scala | package se.parwen.adoc.day17
object Input {
}
|
parwen68/advent-of-code-2017 | build.sbt | enablePlugins(ScalaJSPlugin)
name := "advent-of-code-2017"
version := "1.0"
scalaVersion in ThisBuild := "2.12.4"
lazy val root = project.in(file(".")).
aggregate(adocJS, adocJVM).
settings(
publish := {},
publishLocal := {}
)
lazy val adoc = crossProject.in(file(".")).
settings(
name := "adven... |
parwen68/advent-of-code-2017 | shared/src/main/scala/se/parwen/adoc/day3/Solver.scala | package se.parwen.adoc.day3
object Solver {
val directions: Stream[String] = "R" #:: "U" #:: "L" #:: "D" #:: directions
val multipliers: Stream[Int] = Stream.from(1).flatMap(i => Stream(i, i))
def moves: Stream[Char] = multipliers.zip(directions).flatMap { case (multiplier, direction) => direction * multiplie... |
parwen68/advent-of-code-2017 | shared/src/main/scala/se/parwen/adoc/day18/Input.scala | package se.parwen.adoc.day18
object Input {
val input = """set i 31
|set a 1
|mul p 17
|jgz p p
|mul a 2
|add i -1
|jgz i -2
|add a -1
|set i 127
|set p 735
... |
parwen68/advent-of-code-2017 | shared/src/main/scala/se/parwen/adoc/day15/Solver.scala | <filename>shared/src/main/scala/se/parwen/adoc/day15/Solver.scala
package se.parwen.adoc.day15
object Solver {
def generator(start: Long, factor: Long) : () => Long = {
var prev = start
() => {
val next = prev * factor % 2147483647
prev = next
next
}
}
def doMatch(first: Long, sec... |
parwen68/advent-of-code-2017 | shared/src/main/scala/se/parwen/adoc/day4/Solver.scala | <gh_stars>0
package se.parwen.adoc.day4
object Solver {
def isSame(a: Array[String]): Boolean = {
val b = a.foldLeft(Map[String,Int]())((m,v) => m.updated(v, m.getOrElse(v,0) + 1))
b.values.max == 1
}
def isAnagram(a: Array[String]): Boolean = {
for (w1 <- a) {
for (w2 <- a) {
if (w1 ... |
parwen68/advent-of-code-2017 | shared/src/main/scala/se/parwen/adoc/day2/Adv2.scala | package se.parwen.adoc.day2
object Adv2 {
val inputStr =
"""1224 926 1380 688 845 109 118 88 1275 1306 91 796 102 1361 27 995
|1928 2097 138 1824 198 117 1532 2000 1478 539 1982 125 1856 139 475 1338
|848 202 1116 791 1114 236 183 186 150 1016 1258 84 952 1202 988 866
|946 155 210 980 896 875 9... |
0xmycf/fabric-mod-scala-template | src/main/scala/com/username/modid/language_adapter/CustomLanguageAdapter.scala | <gh_stars>0
package com.username.modid.language_adapter
import net.fabricmc.loader.api.{LanguageAdapter, ModContainer}
class CustomLanguageAdapter extends LanguageAdapter {
override def create[T](mod: ModContainer, value: String, aClass: Class[T]): T =
Class.forName(value + "$").getField("MODULE$").get(null).as... |
0xmycf/fabric-mod-scala-template | src/main/scala/com/username/modid/blocks/ModIdBlocks.scala | package com.username.modid.blocks
import com.username.modid.MainFile
import net.minecraft.block.{AbstractBlock, Block, Blocks}
import net.minecraft.item.{BlockItem, Item, ItemGroup}
import net.minecraft.util.Identifier
import net.minecraft.util.registry.Registry
import scala.collection.mutable
object ModIdBlocks {
... |
0xmycf/fabric-mod-scala-template | src/main/scala/com/username/modid/MainFile.scala | <reponame>0xmycf/fabric-mod-scala-template<filename>src/main/scala/com/username/modid/MainFile.scala
package com.username.modid
import com.username.modid.blocks.ModIdBlocks
import com.username.modid.items.ModIdItems
import net.fabricmc.api.ModInitializer
/**
* Main File
* VM options for mixins:
* -Dmixin.debug.exp... |
0xmycf/fabric-mod-scala-template | src/main/scala/com/username/modid/items/ModIdItems.scala | <reponame>0xmycf/fabric-mod-scala-template
package com.username.modid.items
import com.username.modid.MainFile
import net.minecraft.item.{Item, ItemGroup}
import net.minecraft.util.Identifier
import net.minecraft.util.registry.Registry
object ModIdItems {
private val ItemRegistry = collection.mutable.LinkedHashMap[... |
0xmycf/fabric-mod-scala-template | src/main/scala/com/username/modid/MainFileClient.scala | package com.username.modid
import net.fabricmc.api.{ClientModInitializer, EnvType, Environment}
@Environment(EnvType.CLIENT)
object MainFileClient extends ClientModInitializer {
override def onInitializeClient(): Unit = {
}
}
|
jerneyio/data-weave-native | native-cli/src/test/scala/org/mule/weave/dwnative/cli/DataWeaveCLITest.scala | package org.mule.weave.dwnative.cli
import java.io.ByteArrayInputStream
import java.io.ByteArrayOutputStream
import java.io.PrintStream
import org.scalatest.FreeSpec
import org.scalatest.Matchers
import scala.io.Source
class DataWeaveCLITest extends FreeSpec with Matchers {
"should work with output application/... |
jerneyio/data-weave-native | native-common/src/main/scala/org/mule/weave/dwnative/NativeRuntime.scala | <filename>native-common/src/main/scala/org/mule/weave/dwnative/NativeRuntime.scala
package org.mule.weave.dwnative
import java.io.File
import java.io.OutputStream
import java.io.PrintWriter
import java.io.StringWriter
import org.mule.weave.v2.interpreted.module.WeaveDataFormat
import org.mule.weave.v2.model.ServiceMa... |
jerneyio/data-weave-native | dependency-manager/src/main/scala/org/weave/deps/DependencyAnnotationProcessor.scala | package org.weave.deps
import java.io.File
import java.nio.file.Files
import coursier.MavenRepository
import coursier._
import coursier.cache.ArtifactError
import coursier.cache.Cache
import coursier.core.Authentication
import coursier.util.Gather
import coursier.util.Task
import org.mule.weave.v2.parser.MessageCollec... |
jerneyio/data-weave-native | native-common/src/main/scala/org/mule/weave/dwnative/DataWeaveUtils.scala | package org.mule.weave.dwnative
import java.io.File
import org.mule.weave.v2.env.StaticServiceProvider
import org.mule.weave.v2.env.WeaveRuntime
import org.mule.weave.v2.model.ServiceRegistration
import org.mule.weave.v2.module.DataFormat
import org.mule.weave.v2.module.csv.CSVDataFormat
import org.mule.weave.v2.modu... |
jerneyio/data-weave-native | native-common/src/main/scala/org/mule/weave/dwnative/PathBasedResourceResolver.scala | package org.mule.weave.dwnative
import java.io.File
import java.util.zip.ZipFile
import org.mule.weave.v2.parser.ast.variables.NameIdentifier
import org.mule.weave.v2.sdk.WeaveResource
import org.mule.weave.v2.sdk.WeaveResourceResolver
import org.mule.weave.v2.utils.WeaveConstants
import org.mule.weave.v2.utils.Weave... |
mkunikowski/the-neophytes-guide-to-scala | src/main/scala/parts/part12.scala | package parts
import common.printer.printHeader
/**
* Created by michal on 1/1/15.
* The Neophyte's Guide to Scala Part 12: Type Classes
*/
object part12 extends App {
//The problem
{
printHeader("The problem")
{
object Statistics {
def median(xs: Vector[Double]): Double = xs(xs.size / 2... |
mkunikowski/the-neophytes-guide-to-scala | src/main/scala/parts/part10.scala | <gh_stars>0
package parts
import common.printer.printHeader
/**
* Created by michal on 12/30/14.
* The Neophyte's Guide to Scala Part 10: Staying DRY With Higher-order Functions
*/
object part10 extends App {
// On higher-order functions
// 1. One or more of its parameters is a function, and it returns some value.... |
mkunikowski/the-neophytes-guide-to-scala | src/main/scala/parts/part2.scala | package parts
import common.printer.printHeader
/**
* Created by michal on 12/7/14.
* The Neophyte's Guide to Scala Part 2: Extracting Sequences
*/
object part2 extends App {
{
printHeader("Pattern matches some number of elements")
val xs = 3 :: 6 :: 12 :: Nil
val r = xs match {
... |
mkunikowski/the-neophytes-guide-to-scala | src/main/scala/parts/part3.scala | <reponame>mkunikowski/the-neophytes-guide-to-scala<filename>src/main/scala/parts/part3.scala
package parts
import common.printer.printHeader
/**
* Created by michal on 12/13/14.
* The Neophyte's Guide to Scala Part 3: Patterns Everywhere
*/
object part3 extends App {
//Pattern matching expressions
{
prin... |
mkunikowski/the-neophytes-guide-to-scala | src/main/scala/parts/part5.scala | <reponame>mkunikowski/the-neophytes-guide-to-scala
package parts
import common.printer._
/**
* Created by michal on 12/21/14.
* The Neophyte's Guide to Scala Part 5: The Option Type
*/
object part5 extends App {
case class User(
id: Int,
firstName: String,
... |
mkunikowski/the-neophytes-guide-to-scala | src/main/scala/parts/part11.scala | package parts
import common.printer.printHeader
/**
* Created by michal on 12/30/14.
* The Neophyte's Guide to Scala Part 11: Currying and Partially Applied Functions
*/
object part11 extends App {
//Partially applied functions
{
printHeader("Partially applied functions")
case class Email(
... |
mkunikowski/the-neophytes-guide-to-scala | src/main/scala/parts/part1.scala | <reponame>mkunikowski/the-neophytes-guide-to-scala
package parts
import common.printer.printHeader
/**
* Created by michal on 12/7/14.
* The Neophyte's Guide to Scala Part 1: Extractors
*/
object part1 extends App {
{
printHeader("Extracting list values")
case class User(firstName: String, las... |
mkunikowski/the-neophytes-guide-to-scala | src/main/scala/parts/part13.scala | package parts
import common.printer.printHeader
/**
* Created by michal on 1/2/15.
* The Neophyte's Guide to Scala Part 13: Path-dependent Types
*/
object part13 extends App {
//The problem
{
printHeader("The problem")
object Franchise {
case class Character(name: String)
}
class Franch... |
mkunikowski/the-neophytes-guide-to-scala | src/main/scala/parts/part7.scala | package parts
import common.printer._
/**
* Created by michal on 12/26/14.
* The Neophyte's Guide to Scala Part 7: The Either Type
*/
object part7 extends App {
//Creating an Either
{
printHeader("Creating an Either")
import scala.io.Source
import java.net.URL
def getContent(url: URL): Either[S... |
mkunikowski/the-neophytes-guide-to-scala | src/main/scala/parts/part6.scala | package parts
import common.printer._
/**
* Created by michal on 12/22/14.
* The Neophyte's Guide to Scala Part 6: Error Handling With Try
*/
object part6 extends App {
//Throwing and catching exceptions
{
printHeader("Throwing and catching exceptions")
case class Customer(age: Int)
class Cigaret... |
mkunikowski/the-neophytes-guide-to-scala | build.sbt | name := """the-neophytes-guide-to-scala"""
version := "1.0"
scalaVersion := "2.11.4"
// Change this to another test framework if you prefer
libraryDependencies += "org.scalatest" %% "scalatest" % "2.1.6" % "test"
//Joda time
libraryDependencies += "joda-time" % "joda-time" % "2.1"
libraryDependencies += "org.joda"... |
mkunikowski/the-neophytes-guide-to-scala | src/main/scala/parts/part8.scala | <filename>src/main/scala/parts/part8.scala
package parts
import common.printer.printHeader
/**
* Created by michal on 12/27/14.
* The Neophyte's Guide to Scala Part 8: Welcome to the Future
*/
object part8 extends App {
//Why sequential code can be bad
{
printHeader("Why sequential code can be bad")
i... |
mkunikowski/the-neophytes-guide-to-scala | src/main/scala/parts/part4.scala | package parts
import common.printer.printHeader
/**
* Created by michal on 12/13/14.
* The Neophyte's Guide to Scala Part 4: Pattern Matching Anonymous Functions
*/
object part4 extends App {
{
printHeader("Pattern Matching Anonymous Functions")
val songTitles = List("The White Hare", "Childe the Hunter... |
mkunikowski/the-neophytes-guide-to-scala | src/main/scala/common/printer.scala | package common
/**
* Created by michal on 12/7/14.
*/
package object printer {
def printHeader(header: String) = {
val size = 80
val seperator = "-"
val strHeader = "\n" + seperator * size + "\n" + header + "\n" + seperator * size + "\n"
print(strHeader)
}
}
|
mkunikowski/the-neophytes-guide-to-scala | src/main/scala/parts/part9.scala | package parts
import common.printer.printHeader
import scala.concurrent.Future
/**
* Created by michal on 12/29/14.
* The Neophyte's Guide to Scala Part 9: Promises and Futures in Practice
*/
object part9 extends App {
//Promises
{
printHeader("Promises")
import concurrent.Future
import concurrent... |
akr4/diff-zip | src/main/scala/diff.scala | /*
* Copyright 2012 <NAME>
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in wr... |
akr4/diff-zip | project/build.scala | <gh_stars>1-10
/*
* Copyright 2012 <NAME>
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or ... |
akr4/diff-zip | src/test/scala/diff-test.scala | <gh_stars>1-10
/*
* Copyright 2012 <NAME>
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or ... |
akr4/diff-zip | src/main/scala/zip.scala | <reponame>akr4/diff-zip
/*
* Copyright 2012 <NAME>
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicabl... |
akr4/diff-zip | src/main/scala/main.scala | <filename>src/main/scala/main.scala
/*
* Copyright 2012 <NAME>
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required ... |
Panos-Bletsos/spark-tpcds-benchmark | build.sbt |
name := "spark-tpcds-benchmark"
version := "0.1"
scalaVersion := "2.11.12"
libraryDependencies ++= Seq(
"com.typesafe" % "config" % "1.3.2",
"com.databricks" %% "spark-sql-perf" % "0.5.0-SNAPSHOT",
"org.apache.spark" %% "spark-core" % "2.2.0" % "provided",
"org.apache.spark" %% "spark-sql" % "2.2.0" % "prov... |
Panos-Bletsos/spark-tpcds-benchmark | src/main/scala/bletsos/panos/App.scala | package bletsos.panos
import com.databricks.spark.sql.perf.tpcds.{TPCDS, TPCDSTables}
import com.typesafe.config.{Config, ConfigFactory}
import org.apache.spark.sql.SparkSession
object App {
def main(args: Array[String]): Unit = {
val conf: Config = ConfigFactory.load
val spark: SparkSession = SparkSessi... |
Amab/topias | src/main/scala/editor/CustomFontMetrics.scala | package editor
import java.awt.font.FontRenderContext
import java.awt.{Canvas, Font, FontMetrics, RenderingHints}
import com.intellij.ide.ui.AntialiasingType
import com.intellij.openapi.editor.Editor
import com.intellij.openapi.editor.impl.{EditorImpl, FontInfo}
import com.intellij.util.ui.UIUtil
import javax.swing.p... |
mijicd/declarative-concurrency | example/src/main/scala/lambdaconf/Transfer.scala | <reponame>mijicd/declarative-concurrency
package lambdaconf
import zio._
import zio.clock._
import zio.console._
import zio.duration._
import zio.stm._
/**
* Problem:
* - Alice has 1000$ on her account, while Bob has 0.
* - Bob asks Alice to loan him 10000$.
* - Alice receives 100$ every 10 milliseconds until ... |
mijicd/declarative-concurrency | example/src/main/scala/lambdaconf/Philosophers.scala | <filename>example/src/main/scala/lambdaconf/Philosophers.scala
package lambdaconf
import zio._
import zio.console._
import zio.stm._
object Philosophers extends App {
case object Fork
type Fork = Fork.type
final case class Seat(left: TRef[Option[Fork]], right: TRef[Option[Fork]])
final case class Table(sea... |
mijicd/declarative-concurrency | example/src/main/scala/lambdaconf/PQueue.scala | package lambdaconf
import zio._
import zio.console._
import zio.stm._
object PQueue extends App {
final class PriorityQueue[A](minPriority: TRef[Int], map: TMap[Int, TQueue[A]]) {
def offer(a: A, priority: Int): USTM[Unit] =
for {
queueM <- map.get(priority)
queue <- queueM.fold(TQueue.un... |
mijicd/declarative-concurrency | example/build.sbt | import Dependencies._
inThisBuild(
List(
scalaVersion := "2.13.1",
organization := "com.github.mijicd"
)
)
lazy val root = project
.in(file("."))
.settings(
libraryDependencies += "dev.zio" %% "zio" % "1.0.0-RC21-1"
)
|
emartech/rdb-connector-common | src/test/scala/com/emarsys/rdb/connector/common/models/ConnectorSearchSpec.scala | <gh_stars>0
package com.emarsys.rdb.connector.common.models
import akka.NotUsed
import akka.stream.scaladsl.Source
import com.emarsys.rdb.connector.common.ConnectorResponse
import com.emarsys.rdb.connector.common.models.DataManipulation.Criteria
import com.emarsys.rdb.connector.common.models.DataManipulation.FieldValu... |
emartech/rdb-connector-common | src/test/scala/com/emarsys/rdb/connector/common/defaults/FieldValueConverterSpec.scala | package com.emarsys.rdb.connector.common.defaults
import com.emarsys.rdb.connector.common.defaults.FieldValueConverter._
import com.emarsys.rdb.connector.common.models.DataManipulation.FieldValueWrapper._
import com.emarsys.rdb.connector.common.models.SimpleSelect._
import org.scalatest.{Matchers, WordSpecLike}
class... |
emartech/rdb-connector-common | build.sbt | name := "rdb-connector-common"
version := "0.1-SNAPSHOT"
scalaVersion := "2.12.3"
libraryDependencies ++= {
val scalaTestV = "3.0.1"
Seq(
"com.typesafe.akka" %% "akka-stream" % "2.5.6",
"com.typesafe.akka" %% "akka-stream-testkit" % "2.5.6" % Test,
"org.scalatest" %% "scalatest" % scalaTestV % Test,... |
emartech/rdb-connector-common | src/test/scala/com/emarsys/rdb/connector/common/models/ValidateGroupLimitableQuerySpec.scala | package com.emarsys.rdb.connector.common.models
import com.emarsys.rdb.connector.common.models.SimpleSelect._
import org.scalatest.{Matchers, WordSpecLike}
class ValidateGroupLimitableQuerySpec extends WordSpecLike with Matchers {
import ValidateGroupLimitableQuery.GroupLimitValidationResult._
"ValidateGroupLim... |
emartech/rdb-connector-common | src/main/scala/com/emarsys/rdb/connector/common/models/ValidateDataManipulation.scala | package com.emarsys.rdb.connector.common.models
import com.emarsys.rdb.connector.common.models.DataManipulation.{Criteria, Record, UpdateDefinition}
import com.emarsys.rdb.connector.common.models.Errors.ErrorWithMessage
import com.emarsys.rdb.connector.common.models.ValidateDataManipulation.ValidationResult
import sc... |
emartech/rdb-connector-common | src/main/scala/com/emarsys/rdb/connector/common/defaults/FieldValueConverter.scala | package com.emarsys.rdb.connector.common.defaults
import com.emarsys.rdb.connector.common.models.DataManipulation.FieldValueWrapper
import com.emarsys.rdb.connector.common.models.SimpleSelect.Value
trait FieldValueConverter[T <: FieldValueWrapper] {
def convert(fieldValueWrapper: T): Option[Value]
}
trait FieldVal... |
emartech/rdb-connector-common | src/main/scala/com/emarsys/rdb/connector/common/defaults/GroupWithLimitStage.scala | package com.emarsys.rdb.connector.common.defaults
import akka.NotUsed
import akka.stream.scaladsl.{Flow, Source}
object GroupWithLimitStage {
def apply[K](references: Seq[String], groupLimit: Int): Flow[Seq[String], Seq[String], NotUsed] = {
val upperRefs = references.map(_.toUpperCase)
def groupKey(xs: S... |
emartech/rdb-connector-common | src/main/scala/com/emarsys/rdb/connector/common/models/Errors.scala | package com.emarsys.rdb.connector.common.models
import com.emarsys.rdb.connector.common.models.ValidateDataManipulation.ValidationResult
object Errors {
sealed abstract class ConnectorError(message: String = "") extends Exception(message)
case class ConnectionError(error: Throwable) extends ConnectorError(error... |
emartech/rdb-connector-common | src/main/scala/com/emarsys/rdb/connector/common/models/SimpleSelect.scala | <reponame>emartech/rdb-connector-common
package com.emarsys.rdb.connector.common.models
import com.emarsys.rdb.connector.common.models.SimpleSelect.{Fields, TableName, WhereCondition}
case class SimpleSelect(fields: Fields,
table: TableName,
where: Option[WhereCondition... |
emartech/rdb-connector-common | src/test/scala/com/emarsys/rdb/connector/common/ConnectionConfigSpec.scala | <reponame>emartech/rdb-connector-common
package com.emarsys.rdb.connector.common
import com.emarsys.rdb.connector.common.models.{CommonConnectionReadableData, ConnectionConfig}
import org.scalatest.{Matchers, WordSpecLike}
import spray.json._
class ConnectionConfigSpec extends WordSpecLike with Matchers {
"Connect... |
emartech/rdb-connector-common | src/main/scala/com/emarsys/rdb/connector/common/models/ValidateGroupLimitableQuery.scala | package com.emarsys.rdb.connector.common.models
import com.emarsys.rdb.connector.common.models.SimpleSelect.{And, EqualToValue, Or, WhereCondition}
import com.emarsys.rdb.connector.common.models.ValidateGroupLimitableQuery.GroupLimitValidationResult
trait ValidateGroupLimitableQuery {
import ValidateGroupLimitable... |
emartech/rdb-connector-common | src/test/scala/com/emarsys/rdb/connector/common/models/ValidateDataManipulatorSpec.scala | package com.emarsys.rdb.connector.common.models
import com.emarsys.rdb.connector.common.models.DataManipulation.FieldValueWrapper.StringValue
import com.emarsys.rdb.connector.common.models.DataManipulation.{FieldValueWrapper, UpdateDefinition}
import com.emarsys.rdb.connector.common.models.Errors.TableNotFound
import ... |
emartech/rdb-connector-common | src/main/scala/com/emarsys/rdb/connector/common/models/package.scala | package com.emarsys.rdb.connector.common
package object models {
case class CommonConnectionReadableData(`type`: String, location: String, dataset: String, user: String)
trait ConnectionConfig {
def replica[C <: this.type]: Option[C] = None
def toCommonFormat: CommonConnectionReadableData
final ove... |
emartech/rdb-connector-common | src/test/scala/com/emarsys/rdb/connector/common/models/ConnectorSpec.scala | package com.emarsys.rdb.connector.common.models
import akka.NotUsed
import akka.stream.scaladsl.Source
import com.emarsys.rdb.connector.common
import com.emarsys.rdb.connector.common.ConnectorResponse
import com.emarsys.rdb.connector.common.models.DataManipulation.FieldValueWrapper.StringValue
import com.emarsys.rdb.c... |
emartech/rdb-connector-common | src/test/scala/com/emarsys/rdb/connector/common/models/ConnectorNotImplementedOperationSpec.scala | <reponame>emartech/rdb-connector-common<gh_stars>0
package com.emarsys.rdb.connector.common.models
import com.emarsys.rdb.connector.common.models.DataManipulation.FieldValueWrapper.StringValue
import com.emarsys.rdb.connector.common.models.DataManipulation.UpdateDefinition
import com.emarsys.rdb.connector.common.model... |
emartech/rdb-connector-common | src/main/scala/com/emarsys/rdb/connector/common/package.scala | package com.emarsys.rdb.connector
import com.emarsys.rdb.connector.common.models.Errors.{ConnectorError, NotImplementedOperation}
import scala.concurrent.Future
package object common {
type ConnectorResponse[T] = Future[Either[ConnectorError,T]]
def notImplementedOperation[T](message: String): ConnectorResponse... |
emartech/rdb-connector-common | src/main/scala/com/emarsys/rdb/connector/common/defaults/SqlWriter.scala | <reponame>emartech/rdb-connector-common
package com.emarsys.rdb.connector.common.defaults
import scala.language.implicitConversions
import com.emarsys.rdb.connector.common.models.SimpleSelect
import com.emarsys.rdb.connector.common.models.SimpleSelect._
import scala.annotation.implicitNotFound
@implicitNotFound(msg ... |
emartech/rdb-connector-common | src/main/scala/com/emarsys/rdb/connector/common/models/TableSchemaDescriptors.scala | <filename>src/main/scala/com/emarsys/rdb/connector/common/models/TableSchemaDescriptors.scala
package com.emarsys.rdb.connector.common.models
object TableSchemaDescriptors {
case class TableModel(name: String, isView: Boolean)
case class FieldModel(name: String, columnType: String)
case class FullTableModel(name... |
emartech/rdb-connector-common | src/test/scala/com/emarsys/rdb/connector/common/defaults/DefaultSqlWriterSpec.scala | <filename>src/test/scala/com/emarsys/rdb/connector/common/defaults/DefaultSqlWriterSpec.scala<gh_stars>0
package com.emarsys.rdb.connector.common.defaults
import com.emarsys.rdb.connector.common.models.SimpleSelect._
import org.scalatest.{Matchers, WordSpecLike}
import SqlWriter._
import com.emarsys.rdb.connector.comm... |
emartech/rdb-connector-common | src/main/scala/com/emarsys/rdb/connector/common/models/Connector.scala | package com.emarsys.rdb.connector.common.models
import akka.NotUsed
import akka.stream.scaladsl.Source
import com.emarsys.rdb.connector.common.defaults.{DefaultFieldValueConverters, FieldValueConverters, GroupWithLimitStage}
import com.emarsys.rdb.connector.common.models.DataManipulation.{Criteria, Record, UpdateDefin... |
emartech/rdb-connector-common | src/main/scala/com/emarsys/rdb/connector/common/models/DataManipulation.scala | <gh_stars>0
package com.emarsys.rdb.connector.common.models
object DataManipulation {
type Record = Map[String, FieldValueWrapper]
type Criteria = Record
case class UpdateDefinition(search: Criteria, update: Record)
case class StringedUpdateDefinition(search: Map[String, String], update: Map[String, String]... |
lindemer/pmp | src/test/scala/PmpTester.scala | <filename>src/test/scala/PmpTester.scala<gh_stars>0
/*
* Copyright (c) 2021 <NAME> <<EMAIL>>
*
* SPDX-License-Identifier: Apache-2.0
*/
package pmp
import org.scalatest.FunSuite
import spinal.core._
import spinal.core.sim._
import config.PmpConfig
import scala.sys.process._
import scala.util.Random
class PmpTes... |
lindemer/pmp | src/main/scala/PmpMain.scala | package pmp
import spinal.core._
object PmpMain {
def main(args: Array[String]) {
SpinalConfig(targetDirectory = "rtl").generateVhdl(new PmpController(count = 16))
}
} |
lindemer/pmp | src/main/scala/Pmp.scala | /*
* Copyright (c) 2021 <NAME> <<EMAIL>>
*
* SPDX-License-Identifier: Apache-2.0
*/
package pmp
import spinal.core.sim._
import spinal.core._
import spinal.lib._
/* Each 32-bit pmpcfg# register contains four 8-bit configuration sections.
* These section numbers contain flags which apply to regions defined by th... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.