repo_name
stringlengths
6
97
path
stringlengths
3
341
text
stringlengths
8
1.02M
cquiroz/scalatest
scalactic-test/src/test/scala/org/scalactic/RecursiveConstraintsSpec.scala
<filename>scalactic-test/src/test/scala/org/scalactic/RecursiveConstraintsSpec.scala /* * Copyright 2001-2014 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http:...
cquiroz/scalatest
common-test/src/main/scala/org/scalatest/NyayaGeneratorDrivenPropertyChecks.scala
<reponame>cquiroz/scalatest<filename>common-test/src/main/scala/org/scalatest/NyayaGeneratorDrivenPropertyChecks.scala /* * Copyright 2001-2015 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy...
cquiroz/scalatest
scalactic/src/main/scala/org/scalactic/LowPriorityEqualityConstraints.scala
<reponame>cquiroz/scalatest /* * Copyright 2001-2014 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless requir...
cquiroz/scalatest
scalactic/src/main/scala/org/scalactic/algebra/Monad.scala
<reponame>cquiroz/scalatest<filename>scalactic/src/main/scala/org/scalactic/algebra/Monad.scala<gh_stars>1-10 /* * Copyright 2001-2014 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the L...
cquiroz/scalatest
examples/src/main/scala/org/scalatest/examples/propspec/multi/ExampleSpec.scala
/* * Copyright 2001-2013 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agre...
cquiroz/scalatest
scalatest-test/src/test/scala/org/scalatest/junit/JUnitWrapperSuiteSuite.scala
/* * Copyright 2001-2013 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agre...
cquiroz/scalatest
scalatest/src/main/scala/org/scalatest/tools/ReporterConfiguration.scala
/* * Copyright 2001-2013 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agre...
cquiroz/scalatest
scalactic-macro/src/main/scala/org/scalactic/anyvals/GuessANumberMacros.scala
/* * Copyright 2001-2014 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writ...
cquiroz/scalatest
scalactic-test/src/test/scala/org/scalactic/algebra/MonadSpec.scala
/* * Copyright 2001-2014 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agre...
cquiroz/scalatest
scalactic/src/main/scala/org/scalactic/enablers/Length.scala
<filename>scalactic/src/main/scala/org/scalactic/enablers/Length.scala /* * Copyright 2001-2013 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.o...
cquiroz/scalatest
scalatest/src/main/scala/org/scalatest/AsyncFreeSpecLike.scala
/* * Copyright 2001-2014 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agre...
cquiroz/scalatest
scalatest/src/main/scala/org/scalatest/tools/JUnitXmlReporter.scala
/* * Copyright 2001-2013 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agre...
cquiroz/scalatest
scalactic-test/src/test/scala/org/scalactic/anyvals/PosZDoubleSpec.scala
/* * Copyright 2001-2014 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agre...
cquiroz/scalatest
scalatest/src/main/scala/org/scalatest/exceptions/StackDepthException.scala
/* * Copyright 2001-2013 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agre...
cquiroz/scalatest
scalatest/src/main/scala/org/scalatest/exceptions/TableDrivenPropertyCheckFailedException.scala
/* * Copyright 2001-2013 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agre...
cquiroz/scalatest
scalactic-test/src/test/scala/org/scalactic/algebra/FunctorSpec.scala
<reponame>cquiroz/scalatest<gh_stars>1-10 /* * Copyright 2001-2014 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * *...
cquiroz/scalatest
examples/src/main/scala/org/scalatest/examples/suite/nested/ASCIISuite.scala
<reponame>cquiroz/scalatest /* * Copyright 2001-2013 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless requir...
cquiroz/scalatest
scalactic/src/main/scala/org/scalactic/algebra/Associative.scala
<reponame>cquiroz/scalatest /* * Copyright 2001-2015 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless requir...
cquiroz/scalatest
scalactic/src/main/scala/org/scalactic/enablers/Existence.scala
/* * Copyright 2001-2013 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agre...
cquiroz/scalatest
scalactic/src/main/scala/org/scalactic/EnabledEqualityBetween.scala
/* * Copyright 2001-2014 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agre...
cquiroz/scalatest
scalactic-test/src/test/scala/org/scalactic/LazySeqSpec.scala
<reponame>cquiroz/scalatest /* * Copyright 2001-2015 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless requir...
cquiroz/scalatest
scalactic-test/src/test/scala/org/scalactic/ComplexSpec.scala
<filename>scalactic-test/src/test/scala/org/scalactic/ComplexSpec.scala /* * Copyright 2001-2014 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache....
cquiroz/scalatest
scalactic/src/main/scala/org/scalactic/AsMethods.scala
/* * Copyright 2001-2013 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agre...
cquiroz/scalatest
scalatest/src/main/scala/org/scalatest/laws/MonadLaws.scala
<reponame>cquiroz/scalatest /* * Copyright 2001-2014 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless requir...
cquiroz/scalatest
scalatest-test/src/test/scala/org/scalatest/ShouldBeReadableLogicalAndImplicitSpec.scala
/* * Copyright 2001-2013 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agre...
cquiroz/scalatest
scalatest/src/main/scala/org/scalatest/JavaClassesWrappers.scala
/* * Copyright 2001-2015 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agre...
cquiroz/scalatest
scalatest-test/src/test/scala/org/scalatest/EveryShouldContainSpec.scala
<reponame>cquiroz/scalatest /* * Copyright 2001-2013 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless requir...
cquiroz/scalatest
scalatest-test/src/test/scala/org/scalatest/AllShouldContainElementTypeCheckSpec.scala
/* * Copyright 2001-2013 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agre...
cquiroz/scalatest
scalatest/src/main/scala/org/scalatest/DynaTags.scala
<reponame>cquiroz/scalatest /* * Copyright 2001-2013 Artima, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless requir...
corbantek/play-photostash
build.sbt
name := "play-photostash" version := "1.0" lazy val `play-photostash` = (project in file(".")).enablePlugins(PlayJava) scalaVersion := "2.11.6" //libraryDependencies ++= Seq(javaJdbc, javaEbean, cache, javaWs) libraryDependencies ++= Seq( "com.arangodb" % "arangodb-java-driver" % "2.5.4", "org.imgscalr" % "imgsca...
PerkinElmer/nifi-google-drive-bundle
nifi-google-drive-processor-processors/src/main/scala/com/example/nifi/processors/DownloadFilesProcessor.scala
<gh_stars>1-10 package com.example.nifi.processors import org.apache.nifi.annotation.behavior.{ ReadsAttribute, ReadsAttributes, WritesAttribute, WritesAttributes } import org.apache.nifi.annotation.documentation.{ CapabilityDescription, SeeAlso, Tags } import org.apache.nifi.components.PropertyDescriptor import org.a...
PerkinElmer/nifi-google-drive-bundle
nifi-google-drive-processor-processors/src/main/scala/com/example/nifi/processors/ExampleProcessor.scala
<reponame>PerkinElmer/nifi-google-drive-bundle /* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache Li...
PerkinElmer/nifi-google-drive-bundle
nifi-google-drive-processor-processors/src/main/scala/com/example/nifi/processors/DriveAuth.scala
<reponame>PerkinElmer/nifi-google-drive-bundle package com.example.nifi.processors import com.google.api.client.auth.oauth2.Credential import com.google.api.client.extensions.java6.auth.oauth2.AuthorizationCodeInstalledApp import com.google.api.client.extensions.jetty.auth.oauth2.LocalServerReceiver import com.google....
PerkinElmer/nifi-google-drive-bundle
nifi-google-drive-processor-processors/src/main/scala/com/example/nifi/processors/ListFilesProcessor.scala
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may ...
kabutopia/shadow-stack
src/test/scala/ShadowStackUnitTest.scala
// See README.md for license details. package shadowstack import java.io.File import chisel3.iotesters import chisel3.iotesters.{ChiselFlatSpec, Driver, PeekPokeTester} class ShadowStackUnitTester(c: ShadowStack) extends PeekPokeTester(c) { val retAddList = List(Int) //val dest = rnd.nextInt(0xffffffff) val d...
kabutopia/shadow-stack
src/main/scala/ShadowStack.scala
// See README.md for license details. package shadowstack import chisel3._ /** * * * */ class ShadowStack(depth: Int = 32) extends Module { val io = IO(new Bundle { val curentAddr = Input(UInt(32.W)) val jal = Input(Bool()) val ret = Input(Bool()) val targetAddr = Input...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/broadcast/BroadcastFactory.scala
package org.apache.spark.broadcast import scala.reflect.ClassTag import org.apache.spark.SecurityManager import org.apache.spark.SparkConf /** * An interface for all the broadcast implementations in Spark (to allow * multiple broadcast implementations). SparkContext uses a BroadcastFactory * implementation to in...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/deploy/history/ApplicationHistoryProvider.scala
package org.apache.spark.deploy.history import java.util.concurrent.locks.ReentrantReadWriteLock import java.util.zip.ZipOutputStream import scala.xml.Node import org.apache.spark.SparkException import org.apache.spark.status.api.v1.ApplicationInfo import org.apache.spark.ui.SparkUI /** * A loaded UI for a Spark...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/deploy/master/LeaderElectionAgent.scala
<gh_stars>0 package org.apache.spark.deploy.master import org.apache.spark.annotation.DeveloperApi /** * :: DeveloperApi :: * * A LeaderElectionAgent tracks current master and is a common interface for all election Agents. */ @DeveloperApi trait LeaderElectionAgent { val masterInstance: LeaderElectable def s...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/StatusAPIImpl.scala
<reponame>CrazyTechnology/spark package org.apache.spark private class SparkJobInfoImpl ( val jobId: Int, val stageIds: Array[Int], val status: JobExecutionStatus) extends SparkJobInfo private class SparkStageInfoImpl( val stageId: Int, val currentAttemptId: Int, val submissionTime: Long, ...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/scheduler/cluster/ExecutorInfo.scala
package org.apache.spark.scheduler.cluster import org.apache.spark.annotation.DeveloperApi /** * :: DeveloperApi :: * Stores information about an executor to pass from the scheduler to SparkListeners. */ @DeveloperApi class ExecutorInfo( val executorHost: String, val totalCores: Int, val logUrlMap: Map[St...
CrazyTechnology/spark
examples/src/main/scala/org/apache/spark/examples/sql/SparkSQLExample.scala
package org.apache.spark.examples.sql import org.apache.spark.sql.Row import org.apache.spark.sql.SparkSession import org.apache.spark.sql.types._ object SparkSQLExample { case class Person(name: String, age: Long) def main(args: Array[String]) { val spark = SparkSession .builder().master("local[*]") ...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/memory/MemoryManager.scala
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may ...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/SerializableWritable.scala
<gh_stars>0 package org.apache.spark import java.io._ import org.apache.hadoop.conf.Configuration import org.apache.hadoop.io.ObjectWritable import org.apache.hadoop.io.Writable import org.apache.spark.annotation.DeveloperApi import org.apache.spark.util.Utils @DeveloperApi class SerializableWritable[T <: Writable]...
CrazyTechnology/spark
sql/core/src/main/scala/org/apache/spark/sql/package.scala
package org.apache.spark import org.apache.spark.annotation.{DeveloperApi, Unstable} import org.apache.spark.sql.execution.SparkStrategy /** * Allows the execution of relational queries, including those expressed in SQL using Spark. * 允许执行关系查询,包括使用 Spark 以 SQL 表示的查询。 * @groupname dataType Data types * @groupde...
CrazyTechnology/spark
resource-managers/yarn/src/main/scala/org/apache/spark/scheduler/cluster/YarnScheduler.scala
package org.apache.spark.scheduler.cluster import org.apache.hadoop.yarn.util.RackResolver import org.apache.log4j.{Level, Logger} import org.apache.spark._ import org.apache.spark.scheduler.TaskSchedulerImpl import org.apache.spark.util.Utils private[spark] class YarnScheduler(sc: SparkContext) extends TaskSchedule...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/deploy/master/ExecutorDesc.scala
package org.apache.spark.deploy.master import org.apache.spark.deploy.{ExecutorDescription, ExecutorState} private[master] class ExecutorDesc( val id: Int, val application: ApplicationInfo, val worker: WorkerInfo, val cores: Int, val memory: Int) { var state = ExecutorState.LAUNCHING /** Co...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/deploy/master/MasterSource.scala
package org.apache.spark.deploy.master import com.codahale.metrics.{Gauge, MetricRegistry} import org.apache.spark.metrics.source.Source private[spark] class MasterSource(val master: Master) extends Source { override val metricRegistry = new MetricRegistry() override val sourceName = "master" // Gauge for wo...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/deploy/master/WorkerInfo.scala
<reponame>CrazyTechnology/spark package org.apache.spark.deploy.master import scala.collection.mutable import org.apache.spark.rpc.RpcEndpointRef import org.apache.spark.util.Utils private[spark] class WorkerInfo( val id: String, val host: String, val port: Int, val cores: Int, val memory: Int, ...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/scheduler/TaskDescription.scala
<filename>core/src/main/scala/org/apache/spark/scheduler/TaskDescription.scala<gh_stars>0 /* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF li...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/broadcast/TorrentBroadcastFactory.scala
<filename>core/src/main/scala/org/apache/spark/broadcast/TorrentBroadcastFactory.scala package org.apache.spark.broadcast import scala.reflect.ClassTag import org.apache.spark.{SecurityManager, SparkConf} /** * A [[org.apache.spark.broadcast.Broadcast]] implementation that uses a BitTorrent-like * protocol to do ...
CrazyTechnology/spark
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/codegen/GeneratePredicate.scala
<filename>sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/codegen/GeneratePredicate.scala package org.apache.spark.sql.catalyst.expressions.codegen import org.apache.spark.sql.catalyst.InternalRow import org.apache.spark.sql.catalyst.expressions._ /** * Interface for generated predicate * 生成谓...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/deploy/master/FileSystemPersistenceEngine.scala
<gh_stars>0 package org.apache.spark.deploy.master import java.io._ import scala.reflect.ClassTag import org.apache.spark.internal.Logging import org.apache.spark.serializer.{DeserializationStream, SerializationStream, Serializer} import org.apache.spark.util.Utils /** * Stores data in a single on-disk directory...
CrazyTechnology/spark
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/InternalRow.scala
<gh_stars>0 package org.apache.spark.sql.catalyst import org.apache.spark.sql.catalyst.expressions._ import org.apache.spark.sql.catalyst.util.{ArrayData, MapData} import org.apache.spark.sql.types._ import org.apache.spark.unsafe.types.UTF8String /** * An abstract class for row used internally in Spark SQL, which ...
CrazyTechnology/spark
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DataSourceRDD.scala
package org.apache.spark.sql.execution.datasources.v2 import org.apache.spark._ import org.apache.spark.rdd.RDD import org.apache.spark.sql.catalyst.InternalRow import org.apache.spark.sql.sources.v2.reader.{InputPartition, PartitionReader, PartitionReaderFactory} class DataSourceRDDPartition(val index: Int, val inpu...
CrazyTechnology/spark
resource-managers/yarn/src/main/scala/org/apache/spark/scheduler/cluster/YarnSchedulerBackend.scala
package org.apache.spark.scheduler.cluster import java.util.concurrent.atomic.{AtomicBoolean} import scala.concurrent.ExecutionContext.Implicits.global import scala.concurrent.Future import scala.util.{Failure, Success} import scala.util.control.NonFatal import org.apache.hadoop.yarn.api.records.{ApplicationAttemptI...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/deploy/master/MasterMessages.scala
package org.apache.spark.deploy.master sealed trait MasterMessages extends Serializable /** Contains messages seen only by the Master and its associated entities. */ private[master] object MasterMessages { // LeaderElectionAgent to Master case object ElectedLeader case object RevokedLeadership // Master ...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/deploy/history/HistoryServerArguments.scala
package org.apache.spark.deploy.history import scala.annotation.tailrec import org.apache.spark.SparkConf import org.apache.spark.internal.Logging import org.apache.spark.util.Utils /** * Command-line parser for the [[HistoryServer]]. */ private[history] class HistoryServerArguments(conf: SparkConf, args: Array[S...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/Partitioner.scala
package org.apache.spark import java.io.{IOException, ObjectInputStream, ObjectOutputStream} import scala.collection.mutable import scala.collection.mutable.ArrayBuffer import scala.math.log10 import scala.reflect.ClassTag import scala.util.hashing.byteswap32 import org.apache.spark.rdd.{PartitionPruningRDD, RDD} im...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/deploy/master/DriverInfo.scala
<reponame>CrazyTechnology/spark package org.apache.spark.deploy.master import java.util.Date import org.apache.spark.deploy.DriverDescription import org.apache.spark.util.Utils private[deploy] class DriverInfo( val startTime: Long, val id: String, val desc: DriverDescription, val submitDate: Date) ...
CrazyTechnology/spark
sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/HadoopFsRelationSuite.scala
<gh_stars>0 package org.apache.spark.sql.execution.datasources import java.io.{File, FilenameFilter} import org.apache.spark.sql.QueryTest import org.apache.spark.sql.execution.joins.{BroadcastHashJoinExec, SortMergeJoinExec} import org.apache.spark.sql.test.SharedSQLContext class HadoopFsRelationSuite extends Query...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/deploy/master/RecoveryModeFactory.scala
<reponame>CrazyTechnology/spark package org.apache.spark.deploy.master import org.apache.spark.SparkConf import org.apache.spark.annotation.DeveloperApi import org.apache.spark.internal.Logging import org.apache.spark.serializer.Serializer /** * ::DeveloperApi:: * * Implementation of this class can be plugged in ...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/Partition.scala
package org.apache.spark /** * An identifier for a partition in an RDD. RDD中分区的标识符 */ trait Partition extends Serializable { /** * Get the partition's index within its parent RDD */ def index: Int // A better default implementation of HashCode override def hashCode(): Int = index override def equal...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/deploy/master/RecoveryState.scala
package org.apache.spark.deploy.master private[deploy] object RecoveryState extends Enumeration { type MasterState = Value val STANDBY, ALIVE, RECOVERING, COMPLETING_RECOVERY = Value }
CrazyTechnology/spark
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/json/JsonUtils.scala
package org.apache.spark.sql.execution.datasources.json import org.apache.spark.input.PortableDataStream import org.apache.spark.rdd.RDD import org.apache.spark.sql.Dataset import org.apache.spark.sql.catalyst.json.JSONOptions object JsonUtils { /** * Sample JSON dataset as configured by `samplingRatio`. */ ...
CrazyTechnology/spark
sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveShim.scala
<filename>sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveShim.scala package org.apache.spark.sql.hive import java.io.{InputStream, OutputStream} import java.rmi.server.UID import scala.collection.JavaConverters._ import scala.language.implicitConversions import scala.reflect.ClassTag import com.google.commo...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/SparkException.scala
<gh_stars>0 package org.apache.spark class SparkException(message: String, cause: Throwable) extends Exception(message, cause) { def this(message: String) = this(message, null) } /** * Exception thrown when execution of some user code in the driver process fails, e.g. * accumulator update fails or failure in t...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/rpc/RpcEndpointAddress.scala
<filename>core/src/main/scala/org/apache/spark/rpc/RpcEndpointAddress.scala package org.apache.spark.rpc import org.apache.spark.SparkException /** * An address identifier for an RPC endpoint. * * The `rpcAddress` may be null, in which case the endpoint is registered via a client-only * connection and can only be...
CrazyTechnology/spark
sql/catalyst/src/main/scala/org/apache/spark/sql/internal/ReadOnlySQLConf.scala
package org.apache.spark.sql.internal import java.util.{Map => JMap} import org.apache.spark.TaskContext import org.apache.spark.internal.config.{ConfigEntry, ConfigProvider, ConfigReader} /** * A readonly SQLConf that will be created by tasks running at the executor side. It reads the * configs from the local pro...
CrazyTechnology/spark
core/src/test/scala/org/apache/spark/scheduler/ExternalClusterManagerSuite.scala
package org.apache.spark.scheduler import org.apache.spark.{LocalSparkContext, SparkConf, SparkContext, SparkFunSuite} import org.apache.spark.executor.ExecutorMetrics import org.apache.spark.scheduler.SchedulingMode.SchedulingMode import org.apache.spark.storage.BlockManagerId import org.apache.spark.util.Accumulato...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/TaskState.scala
<reponame>CrazyTechnology/spark package org.apache.spark private[spark] object TaskState extends Enumeration { val LAUNCHING, RUNNING, FINISHED, FAILED, KILLED, LOST = Value private val FINISHED_STATES = Set(FINISHED, FAILED, KILLED, LOST) type TaskState = Value def isFailed(state: TaskState): Boolean = (...
CrazyTechnology/spark
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/plans/logical/Command.scala
<gh_stars>0 package org.apache.spark.sql.catalyst.plans.logical import org.apache.spark.sql.catalyst.expressions.Attribute /** * A logical node that represents a non-query command to be executed by the system. For example, * commands can be used by parsers to represent DDL operations. Commands, unlike queries, a...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/deploy/master/DriverState.scala
<reponame>CrazyTechnology/spark package org.apache.spark.deploy.master private[deploy] object DriverState extends Enumeration { type DriverState = Value // SUBMITTED: Submitted but not yet scheduled on a worker // RUNNING: Has been allocated to a worker to run // FINISHED: Previously ran and exited cleanly ...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/deploy/history/FsHistoryProvider.scala
package org.apache.spark.deploy.history import java.io.{File, FileNotFoundException, IOException} import java.nio.file.Files import java.util.{Date, ServiceLoader} import java.util.concurrent.{ConcurrentHashMap, ExecutorService, Future, TimeUnit} import java.util.zip.{ZipEntry, ZipOutputStream} import scala.collecti...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/Aggregator.scala
<gh_stars>0 package org.apache.spark import org.apache.spark.annotation.DeveloperApi import org.apache.spark.util.collection.ExternalAppendOnlyMap /** * :: DeveloperApi :: * A set of functions used to aggregate data. * * @param createCombiner function to create the initial value of the aggregation. * @param merg...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/deploy/master/ZooKeeperLeaderElectionAgent.scala
package org.apache.spark.deploy.master import org.apache.curator.framework.CuratorFramework import org.apache.curator.framework.recipes.leader.{LeaderLatch, LeaderLatchListener} import org.apache.spark.SparkConf import org.apache.spark.deploy.SparkCuratorUtil import org.apache.spark.internal.Logging private[master]...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/deploy/master/WorkerState.scala
<filename>core/src/main/scala/org/apache/spark/deploy/master/WorkerState.scala<gh_stars>0 package org.apache.spark.deploy.master private[master] object WorkerState extends Enumeration { type WorkerState = Value val ALIVE, DEAD, DECOMMISSIONED, UNKNOWN = Value }
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/scheduler/DAGSchedulerSource.scala
<filename>core/src/main/scala/org/apache/spark/scheduler/DAGSchedulerSource.scala package org.apache.spark.scheduler import com.codahale.metrics.{Gauge, MetricRegistry, Timer} import org.apache.spark.metrics.source.Source private[scheduler] class DAGSchedulerSource(val dagScheduler: DAGScheduler) extends Source ...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/util/taskListeners.scala
package org.apache.spark.util import java.util.EventListener import org.apache.spark.TaskContext import org.apache.spark.annotation.DeveloperApi /** * :: DeveloperApi :: * * Listener providing a callback function to invoke when a task's execution completes. */ @DeveloperApi trait TaskCompletionListener extends ...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/deploy/master/ZooKeeperPersistenceEngine.scala
<gh_stars>0 package org.apache.spark.deploy.master import java.nio.ByteBuffer import scala.collection.JavaConverters._ import scala.reflect.ClassTag import org.apache.curator.framework.CuratorFramework import org.apache.zookeeper.CreateMode import org.apache.spark.SparkConf import org.apache.spark.deploy.SparkCura...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/scheduler/SchedulerBackend.scala
<reponame>CrazyTechnology/spark<filename>core/src/main/scala/org/apache/spark/scheduler/SchedulerBackend.scala package org.apache.spark.scheduler /** * A backend interface for scheduling systems that allows plugging in different ones under * TaskSchedulerImpl. We assume a Mesos-like model where the application gets ...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/broadcast/BroadcastManager.scala
<reponame>CrazyTechnology/spark package org.apache.spark.broadcast import java.util.concurrent.atomic.AtomicLong import scala.reflect.ClassTag import org.apache.commons.collections.map.{AbstractReferenceMap, ReferenceMap} import org.apache.spark.{SecurityManager, SparkConf} import org.apache.spark.internal.Logging...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/deploy/master/ApplicationState.scala
package org.apache.spark.deploy.master private[master] object ApplicationState extends Enumeration { type ApplicationState = Value val WAITING, RUNNING, FINISHED, FAILED, KILLED, UNKNOWN = Value }
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/metrics/source/Source.scala
<reponame>CrazyTechnology/spark package org.apache.spark.metrics.source import com.codahale.metrics.MetricRegistry private[spark] trait Source { def sourceName: String def metricRegistry: MetricRegistry }
CrazyTechnology/spark
sql/hive/src/main/scala/org/apache/spark/sql/hive/package.scala
<gh_stars>0 package org.apache.spark.sql /** * Support for running Spark SQL queries using functionality from Apache Hive (does not require an * existing Hive installation). Supported Hive features include: * - Using HiveQL to express queries. * - Reading metadata from the Hive Metastore using HiveSerDes. * ...
CrazyTechnology/spark
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DataSourceV2Relation.scala
<reponame>CrazyTechnology/spark package org.apache.spark.sql.execution.datasources.v2 import java.util.UUID import scala.collection.JavaConverters._ import org.apache.spark.sql.{AnalysisException, SaveMode} import org.apache.spark.sql.catalyst.analysis.{MultiInstanceRelation, NamedRelation} import org.apache.spark....
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/deploy/client/StandaloneAppClientListener.scala
package org.apache.spark.deploy.client /** * Callbacks invoked by deploy client when various events happen. There are currently five events: * connecting to the cluster, disconnecting, being given an executor, having an executor removed * (either due to failure or due to revocation), and having a worker removed. ...
CrazyTechnology/spark
core/src/main/scala/org/apache/spark/scheduler/cluster/ExecutorData.scala
<gh_stars>0 package org.apache.spark.scheduler.cluster import org.apache.spark.rpc.{RpcAddress, RpcEndpointRef} /** * Grouping of data for an executor used by CoarseGrainedSchedulerBackend. * * @param executorEndpoint The RpcEndpointRef representing this executor * @param executorAddress The network address of t...
lamastex/spark-trend-calculus-examples
notebooks/db/03streamable-trend-calculus-estimators.scala
// Databricks notebook source // MAGIC %md // MAGIC # Markov Model for Trend Calculus // MAGIC // MAGIC <NAME>, <NAME> and <NAME> // MAGIC // MAGIC 2020, Uppsala, Sweden // MAGIC // MAGIC This project was supported by Combient Mix AB through summer internships at: // MAGIC // MAGIC Combient Competence Centre for Da...
lamastex/spark-trend-calculus-examples
notebooks/db/gdelt-POI-detection.scala
<reponame>lamastex/spark-trend-calculus-examples // Databricks notebook source // MAGIC %md // MAGIC # Detecting Persons of Interest to OIL/GAS Price Trends // MAGIC // MAGIC <NAME>, <NAME> and <NAME> // MAGIC // MAGIC 2020, Uppsala, Sweden // MAGIC // MAGIC // MAGIC This project was supported by Combient Mix AB th...
lamastex/spark-trend-calculus-examples
notebooks/db/FX1M.scala
// Databricks notebook source // MAGIC %md // MAGIC # Historical Yahoo! Finance data // MAGIC // MAGIC <NAME>, <NAME> and <NAME> // MAGIC // MAGIC 2020, Uppsala, Sweden // MAGIC // MAGIC This project was supported by Combient Mix AB through summer internships at: // MAGIC // MAGIC Combient Competence Centre for Dat...
lamastex/spark-trend-calculus-examples
notebooks/db/Overview.scala
<reponame>lamastex/spark-trend-calculus-examples // Databricks notebook source // MAGIC %md // MAGIC # Trend Calculus of OIL Price // MAGIC // MAGIC <NAME>, <NAME> and <NAME> // MAGIC // MAGIC 2020, Uppsala, Sweden // MAGIC // MAGIC This work was inspired by <NAME>'s texata-2017 repository forked here: // MAGIC // ...
lamastex/spark-trend-calculus-examples
notebooks/db/spark-gdelt-examples/gdelt-EOI-detection.scala
<reponame>lamastex/spark-trend-calculus-examples<filename>notebooks/db/spark-gdelt-examples/gdelt-EOI-detection.scala // Databricks notebook source // MAGIC %md // MAGIC # Detecting Events of Interest to OIL/GAS Price Trends // MAGIC // MAGIC <NAME>, <NAME> and <NAME> // MAGIC // MAGIC 2020, Uppsala, Sweden // MAGIC ...
atomicbits/scramlgen
modules/scraml-raml-parser/src/main/scala/io/atomicbits/scraml/ramlparser/parser/KeyedList.scala
/* * * (C) Copyright 2018 Atomic BITS (http://atomicbits.io). * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by appl...
atomicbits/scramlgen
modules/scraml-dsl-scala/src/main/scala/io/atomicbits/scraml/dsl/scalaplay/client/ClientConfig.scala
/* * * (C) Copyright 2018 Atomic BITS (http://atomicbits.io). * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by appl...
atomicbits/scramlgen
modules/scraml-raml-parser/src/main/scala/io/atomicbits/scraml/ramlparser/model/parsedtypes/ParsedGenericObject.scala
<reponame>atomicbits/scramlgen /* * * (C) Copyright 2018 Atomic BITS (http://atomicbits.io). * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * http://www.apache.org/licenses/LICENSE-2....
atomicbits/scramlgen
modules/scraml-raml-parser/src/test/scala/io/atomicbits/scraml/ramlparser/model/FetchReplaceStringsTest.scala
<filename>modules/scraml-raml-parser/src/test/scala/io/atomicbits/scraml/ramlparser/model/FetchReplaceStringsTest.scala /* * * (C) Copyright 2018 Atomic BITS (http://atomicbits.io). * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License...
atomicbits/scramlgen
modules/scraml-generator/src/main/scala/io/atomicbits/scraml/generator/platform/javajackson/InterfaceGenerator.scala
/* * * (C) Copyright 2018 Atomic BITS (http://atomicbits.io). * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by appl...
atomicbits/scramlgen
modules/scraml-raml-parser/src/main/scala/io/atomicbits/scraml/ramlparser/lookup/transformers/FallbackTransformer.scala
/* * * (C) Copyright 2018 Atomic BITS (http://atomicbits.io). * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by appl...
atomicbits/scramlgen
modules/scraml-generator/src/test/scala/io/atomicbits/scraml/generator/WithEnumGeneratorTest.scala
/* * * (C) Copyright 2018 Atomic BITS (http://atomicbits.io). * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by appl...