repo_name stringlengths 6 97 | path stringlengths 3 341 | text stringlengths 8 1.02M |
|---|---|---|
cquiroz/scalatest | scalactic-test/src/test/scala/org/scalactic/RecursiveConstraintsSpec.scala | <filename>scalactic-test/src/test/scala/org/scalactic/RecursiveConstraintsSpec.scala
/*
* Copyright 2001-2014 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http:... |
cquiroz/scalatest | common-test/src/main/scala/org/scalatest/NyayaGeneratorDrivenPropertyChecks.scala | <reponame>cquiroz/scalatest<filename>common-test/src/main/scala/org/scalatest/NyayaGeneratorDrivenPropertyChecks.scala
/*
* Copyright 2001-2015 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy... |
cquiroz/scalatest | scalactic/src/main/scala/org/scalactic/LowPriorityEqualityConstraints.scala | <reponame>cquiroz/scalatest
/*
* Copyright 2001-2014 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless requir... |
cquiroz/scalatest | scalactic/src/main/scala/org/scalactic/algebra/Monad.scala | <reponame>cquiroz/scalatest<filename>scalactic/src/main/scala/org/scalactic/algebra/Monad.scala<gh_stars>1-10
/*
* Copyright 2001-2014 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the L... |
cquiroz/scalatest | examples/src/main/scala/org/scalatest/examples/propspec/multi/ExampleSpec.scala | /*
* Copyright 2001-2013 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agre... |
cquiroz/scalatest | scalatest-test/src/test/scala/org/scalatest/junit/JUnitWrapperSuiteSuite.scala | /*
* Copyright 2001-2013 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agre... |
cquiroz/scalatest | scalatest/src/main/scala/org/scalatest/tools/ReporterConfiguration.scala | /*
* Copyright 2001-2013 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agre... |
cquiroz/scalatest | scalactic-macro/src/main/scala/org/scalactic/anyvals/GuessANumberMacros.scala | /*
* Copyright 2001-2014 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writ... |
cquiroz/scalatest | scalactic-test/src/test/scala/org/scalactic/algebra/MonadSpec.scala | /*
* Copyright 2001-2014 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agre... |
cquiroz/scalatest | scalactic/src/main/scala/org/scalactic/enablers/Length.scala | <filename>scalactic/src/main/scala/org/scalactic/enablers/Length.scala
/*
* Copyright 2001-2013 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.o... |
cquiroz/scalatest | scalatest/src/main/scala/org/scalatest/AsyncFreeSpecLike.scala | /*
* Copyright 2001-2014 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agre... |
cquiroz/scalatest | scalatest/src/main/scala/org/scalatest/tools/JUnitXmlReporter.scala | /*
* Copyright 2001-2013 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agre... |
cquiroz/scalatest | scalactic-test/src/test/scala/org/scalactic/anyvals/PosZDoubleSpec.scala | /*
* Copyright 2001-2014 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agre... |
cquiroz/scalatest | scalatest/src/main/scala/org/scalatest/exceptions/StackDepthException.scala | /*
* Copyright 2001-2013 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agre... |
cquiroz/scalatest | scalatest/src/main/scala/org/scalatest/exceptions/TableDrivenPropertyCheckFailedException.scala | /*
* Copyright 2001-2013 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agre... |
cquiroz/scalatest | scalactic-test/src/test/scala/org/scalactic/algebra/FunctorSpec.scala | <reponame>cquiroz/scalatest<gh_stars>1-10
/*
* Copyright 2001-2014 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
*... |
cquiroz/scalatest | examples/src/main/scala/org/scalatest/examples/suite/nested/ASCIISuite.scala | <reponame>cquiroz/scalatest
/*
* Copyright 2001-2013 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless requir... |
cquiroz/scalatest | scalactic/src/main/scala/org/scalactic/algebra/Associative.scala | <reponame>cquiroz/scalatest
/*
* Copyright 2001-2015 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless requir... |
cquiroz/scalatest | scalactic/src/main/scala/org/scalactic/enablers/Existence.scala | /*
* Copyright 2001-2013 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agre... |
cquiroz/scalatest | scalactic/src/main/scala/org/scalactic/EnabledEqualityBetween.scala | /*
* Copyright 2001-2014 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agre... |
cquiroz/scalatest | scalactic-test/src/test/scala/org/scalactic/LazySeqSpec.scala | <reponame>cquiroz/scalatest
/*
* Copyright 2001-2015 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless requir... |
cquiroz/scalatest | scalactic-test/src/test/scala/org/scalactic/ComplexSpec.scala | <filename>scalactic-test/src/test/scala/org/scalactic/ComplexSpec.scala
/*
* Copyright 2001-2014 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.... |
cquiroz/scalatest | scalactic/src/main/scala/org/scalactic/AsMethods.scala | /*
* Copyright 2001-2013 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agre... |
cquiroz/scalatest | scalatest/src/main/scala/org/scalatest/laws/MonadLaws.scala | <reponame>cquiroz/scalatest
/*
* Copyright 2001-2014 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless requir... |
cquiroz/scalatest | scalatest-test/src/test/scala/org/scalatest/ShouldBeReadableLogicalAndImplicitSpec.scala | /*
* Copyright 2001-2013 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agre... |
cquiroz/scalatest | scalatest/src/main/scala/org/scalatest/JavaClassesWrappers.scala | /*
* Copyright 2001-2015 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agre... |
cquiroz/scalatest | scalatest-test/src/test/scala/org/scalatest/EveryShouldContainSpec.scala | <reponame>cquiroz/scalatest
/*
* Copyright 2001-2013 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless requir... |
cquiroz/scalatest | scalatest-test/src/test/scala/org/scalatest/AllShouldContainElementTypeCheckSpec.scala | /*
* Copyright 2001-2013 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agre... |
cquiroz/scalatest | scalatest/src/main/scala/org/scalatest/DynaTags.scala | <reponame>cquiroz/scalatest
/*
* Copyright 2001-2013 Artima, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless requir... |
corbantek/play-photostash | build.sbt | name := "play-photostash"
version := "1.0"
lazy val `play-photostash` = (project in file(".")).enablePlugins(PlayJava)
scalaVersion := "2.11.6"
//libraryDependencies ++= Seq(javaJdbc, javaEbean, cache, javaWs)
libraryDependencies ++= Seq(
"com.arangodb" % "arangodb-java-driver" % "2.5.4",
"org.imgscalr" % "imgsca... |
PerkinElmer/nifi-google-drive-bundle | nifi-google-drive-processor-processors/src/main/scala/com/example/nifi/processors/DownloadFilesProcessor.scala | <gh_stars>1-10
package com.example.nifi.processors
import org.apache.nifi.annotation.behavior.{ ReadsAttribute, ReadsAttributes, WritesAttribute, WritesAttributes }
import org.apache.nifi.annotation.documentation.{ CapabilityDescription, SeeAlso, Tags }
import org.apache.nifi.components.PropertyDescriptor
import org.a... |
PerkinElmer/nifi-google-drive-bundle | nifi-google-drive-processor-processors/src/main/scala/com/example/nifi/processors/ExampleProcessor.scala | <reponame>PerkinElmer/nifi-google-drive-bundle
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache Li... |
PerkinElmer/nifi-google-drive-bundle | nifi-google-drive-processor-processors/src/main/scala/com/example/nifi/processors/DriveAuth.scala | <reponame>PerkinElmer/nifi-google-drive-bundle
package com.example.nifi.processors
import com.google.api.client.auth.oauth2.Credential
import com.google.api.client.extensions.java6.auth.oauth2.AuthorizationCodeInstalledApp
import com.google.api.client.extensions.jetty.auth.oauth2.LocalServerReceiver
import com.google.... |
PerkinElmer/nifi-google-drive-bundle | nifi-google-drive-processor-processors/src/main/scala/com/example/nifi/processors/ListFilesProcessor.scala | /*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may ... |
kabutopia/shadow-stack | src/test/scala/ShadowStackUnitTest.scala | // See README.md for license details.
package shadowstack
import java.io.File
import chisel3.iotesters
import chisel3.iotesters.{ChiselFlatSpec, Driver, PeekPokeTester}
class ShadowStackUnitTester(c: ShadowStack) extends PeekPokeTester(c) {
val retAddList = List(Int)
//val dest = rnd.nextInt(0xffffffff)
val d... |
kabutopia/shadow-stack | src/main/scala/ShadowStack.scala | // See README.md for license details.
package shadowstack
import chisel3._
/**
*
*
*
*/
class ShadowStack(depth: Int = 32) extends Module {
val io = IO(new Bundle {
val curentAddr = Input(UInt(32.W))
val jal = Input(Bool())
val ret = Input(Bool())
val targetAddr = Input... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/broadcast/BroadcastFactory.scala |
package org.apache.spark.broadcast
import scala.reflect.ClassTag
import org.apache.spark.SecurityManager
import org.apache.spark.SparkConf
/**
* An interface for all the broadcast implementations in Spark (to allow
* multiple broadcast implementations). SparkContext uses a BroadcastFactory
* implementation to in... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/deploy/history/ApplicationHistoryProvider.scala |
package org.apache.spark.deploy.history
import java.util.concurrent.locks.ReentrantReadWriteLock
import java.util.zip.ZipOutputStream
import scala.xml.Node
import org.apache.spark.SparkException
import org.apache.spark.status.api.v1.ApplicationInfo
import org.apache.spark.ui.SparkUI
/**
* A loaded UI for a Spark... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/deploy/master/LeaderElectionAgent.scala | <gh_stars>0
package org.apache.spark.deploy.master
import org.apache.spark.annotation.DeveloperApi
/**
* :: DeveloperApi ::
*
* A LeaderElectionAgent tracks current master and is a common interface for all election Agents.
*/
@DeveloperApi
trait LeaderElectionAgent {
val masterInstance: LeaderElectable
def s... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/StatusAPIImpl.scala | <reponame>CrazyTechnology/spark
package org.apache.spark
private class SparkJobInfoImpl (
val jobId: Int,
val stageIds: Array[Int],
val status: JobExecutionStatus)
extends SparkJobInfo
private class SparkStageInfoImpl(
val stageId: Int,
val currentAttemptId: Int,
val submissionTime: Long,
... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/scheduler/cluster/ExecutorInfo.scala | package org.apache.spark.scheduler.cluster
import org.apache.spark.annotation.DeveloperApi
/**
* :: DeveloperApi ::
* Stores information about an executor to pass from the scheduler to SparkListeners.
*/
@DeveloperApi
class ExecutorInfo(
val executorHost: String,
val totalCores: Int,
val logUrlMap: Map[St... |
CrazyTechnology/spark | examples/src/main/scala/org/apache/spark/examples/sql/SparkSQLExample.scala | package org.apache.spark.examples.sql
import org.apache.spark.sql.Row
import org.apache.spark.sql.SparkSession
import org.apache.spark.sql.types._
object SparkSQLExample {
case class Person(name: String, age: Long)
def main(args: Array[String]) {
val spark = SparkSession
.builder().master("local[*]")
... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/memory/MemoryManager.scala | /*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may ... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/SerializableWritable.scala | <gh_stars>0
package org.apache.spark
import java.io._
import org.apache.hadoop.conf.Configuration
import org.apache.hadoop.io.ObjectWritable
import org.apache.hadoop.io.Writable
import org.apache.spark.annotation.DeveloperApi
import org.apache.spark.util.Utils
@DeveloperApi
class SerializableWritable[T <: Writable]... |
CrazyTechnology/spark | sql/core/src/main/scala/org/apache/spark/sql/package.scala |
package org.apache.spark
import org.apache.spark.annotation.{DeveloperApi, Unstable}
import org.apache.spark.sql.execution.SparkStrategy
/**
* Allows the execution of relational queries, including those expressed in SQL using Spark.
* 允许执行关系查询,包括使用 Spark 以 SQL 表示的查询。
* @groupname dataType Data types
* @groupde... |
CrazyTechnology/spark | resource-managers/yarn/src/main/scala/org/apache/spark/scheduler/cluster/YarnScheduler.scala | package org.apache.spark.scheduler.cluster
import org.apache.hadoop.yarn.util.RackResolver
import org.apache.log4j.{Level, Logger}
import org.apache.spark._
import org.apache.spark.scheduler.TaskSchedulerImpl
import org.apache.spark.util.Utils
private[spark] class YarnScheduler(sc: SparkContext) extends TaskSchedule... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/deploy/master/ExecutorDesc.scala |
package org.apache.spark.deploy.master
import org.apache.spark.deploy.{ExecutorDescription, ExecutorState}
private[master] class ExecutorDesc(
val id: Int,
val application: ApplicationInfo,
val worker: WorkerInfo,
val cores: Int,
val memory: Int) {
var state = ExecutorState.LAUNCHING
/** Co... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/deploy/master/MasterSource.scala |
package org.apache.spark.deploy.master
import com.codahale.metrics.{Gauge, MetricRegistry}
import org.apache.spark.metrics.source.Source
private[spark] class MasterSource(val master: Master) extends Source {
override val metricRegistry = new MetricRegistry()
override val sourceName = "master"
// Gauge for wo... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/deploy/master/WorkerInfo.scala | <reponame>CrazyTechnology/spark
package org.apache.spark.deploy.master
import scala.collection.mutable
import org.apache.spark.rpc.RpcEndpointRef
import org.apache.spark.util.Utils
private[spark] class WorkerInfo(
val id: String,
val host: String,
val port: Int,
val cores: Int,
val memory: Int,
... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/scheduler/TaskDescription.scala | <filename>core/src/main/scala/org/apache/spark/scheduler/TaskDescription.scala<gh_stars>0
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF li... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/broadcast/TorrentBroadcastFactory.scala | <filename>core/src/main/scala/org/apache/spark/broadcast/TorrentBroadcastFactory.scala
package org.apache.spark.broadcast
import scala.reflect.ClassTag
import org.apache.spark.{SecurityManager, SparkConf}
/**
* A [[org.apache.spark.broadcast.Broadcast]] implementation that uses a BitTorrent-like
* protocol to do ... |
CrazyTechnology/spark | sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/codegen/GeneratePredicate.scala | <filename>sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/codegen/GeneratePredicate.scala
package org.apache.spark.sql.catalyst.expressions.codegen
import org.apache.spark.sql.catalyst.InternalRow
import org.apache.spark.sql.catalyst.expressions._
/**
* Interface for generated predicate
* 生成谓... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/deploy/master/FileSystemPersistenceEngine.scala | <gh_stars>0
package org.apache.spark.deploy.master
import java.io._
import scala.reflect.ClassTag
import org.apache.spark.internal.Logging
import org.apache.spark.serializer.{DeserializationStream, SerializationStream, Serializer}
import org.apache.spark.util.Utils
/**
* Stores data in a single on-disk directory... |
CrazyTechnology/spark | sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/InternalRow.scala | <gh_stars>0
package org.apache.spark.sql.catalyst
import org.apache.spark.sql.catalyst.expressions._
import org.apache.spark.sql.catalyst.util.{ArrayData, MapData}
import org.apache.spark.sql.types._
import org.apache.spark.unsafe.types.UTF8String
/**
* An abstract class for row used internally in Spark SQL, which ... |
CrazyTechnology/spark | sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DataSourceRDD.scala | package org.apache.spark.sql.execution.datasources.v2
import org.apache.spark._
import org.apache.spark.rdd.RDD
import org.apache.spark.sql.catalyst.InternalRow
import org.apache.spark.sql.sources.v2.reader.{InputPartition, PartitionReader, PartitionReaderFactory}
class DataSourceRDDPartition(val index: Int, val inpu... |
CrazyTechnology/spark | resource-managers/yarn/src/main/scala/org/apache/spark/scheduler/cluster/YarnSchedulerBackend.scala | package org.apache.spark.scheduler.cluster
import java.util.concurrent.atomic.{AtomicBoolean}
import scala.concurrent.ExecutionContext.Implicits.global
import scala.concurrent.Future
import scala.util.{Failure, Success}
import scala.util.control.NonFatal
import org.apache.hadoop.yarn.api.records.{ApplicationAttemptI... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/deploy/master/MasterMessages.scala |
package org.apache.spark.deploy.master
sealed trait MasterMessages extends Serializable
/** Contains messages seen only by the Master and its associated entities. */
private[master] object MasterMessages {
// LeaderElectionAgent to Master
case object ElectedLeader
case object RevokedLeadership
// Master ... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/deploy/history/HistoryServerArguments.scala |
package org.apache.spark.deploy.history
import scala.annotation.tailrec
import org.apache.spark.SparkConf
import org.apache.spark.internal.Logging
import org.apache.spark.util.Utils
/**
* Command-line parser for the [[HistoryServer]].
*/
private[history] class HistoryServerArguments(conf: SparkConf, args: Array[S... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/Partitioner.scala | package org.apache.spark
import java.io.{IOException, ObjectInputStream, ObjectOutputStream}
import scala.collection.mutable
import scala.collection.mutable.ArrayBuffer
import scala.math.log10
import scala.reflect.ClassTag
import scala.util.hashing.byteswap32
import org.apache.spark.rdd.{PartitionPruningRDD, RDD}
im... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/deploy/master/DriverInfo.scala | <reponame>CrazyTechnology/spark
package org.apache.spark.deploy.master
import java.util.Date
import org.apache.spark.deploy.DriverDescription
import org.apache.spark.util.Utils
private[deploy] class DriverInfo(
val startTime: Long,
val id: String,
val desc: DriverDescription,
val submitDate: Date)
... |
CrazyTechnology/spark | sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/HadoopFsRelationSuite.scala | <gh_stars>0
package org.apache.spark.sql.execution.datasources
import java.io.{File, FilenameFilter}
import org.apache.spark.sql.QueryTest
import org.apache.spark.sql.execution.joins.{BroadcastHashJoinExec, SortMergeJoinExec}
import org.apache.spark.sql.test.SharedSQLContext
class HadoopFsRelationSuite extends Query... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/deploy/master/RecoveryModeFactory.scala | <reponame>CrazyTechnology/spark
package org.apache.spark.deploy.master
import org.apache.spark.SparkConf
import org.apache.spark.annotation.DeveloperApi
import org.apache.spark.internal.Logging
import org.apache.spark.serializer.Serializer
/**
* ::DeveloperApi::
*
* Implementation of this class can be plugged in ... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/Partition.scala | package org.apache.spark
/**
* An identifier for a partition in an RDD. RDD中分区的标识符
*/
trait Partition extends Serializable {
/**
* Get the partition's index within its parent RDD
*/
def index: Int
// A better default implementation of HashCode
override def hashCode(): Int = index
override def equal... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/deploy/master/RecoveryState.scala |
package org.apache.spark.deploy.master
private[deploy] object RecoveryState extends Enumeration {
type MasterState = Value
val STANDBY, ALIVE, RECOVERING, COMPLETING_RECOVERY = Value
}
|
CrazyTechnology/spark | sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/json/JsonUtils.scala |
package org.apache.spark.sql.execution.datasources.json
import org.apache.spark.input.PortableDataStream
import org.apache.spark.rdd.RDD
import org.apache.spark.sql.Dataset
import org.apache.spark.sql.catalyst.json.JSONOptions
object JsonUtils {
/**
* Sample JSON dataset as configured by `samplingRatio`.
*/
... |
CrazyTechnology/spark | sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveShim.scala | <filename>sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveShim.scala
package org.apache.spark.sql.hive
import java.io.{InputStream, OutputStream}
import java.rmi.server.UID
import scala.collection.JavaConverters._
import scala.language.implicitConversions
import scala.reflect.ClassTag
import com.google.commo... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/SparkException.scala | <gh_stars>0
package org.apache.spark
class SparkException(message: String, cause: Throwable)
extends Exception(message, cause) {
def this(message: String) = this(message, null)
}
/**
* Exception thrown when execution of some user code in the driver process fails, e.g.
* accumulator update fails or failure in t... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/rpc/RpcEndpointAddress.scala | <filename>core/src/main/scala/org/apache/spark/rpc/RpcEndpointAddress.scala
package org.apache.spark.rpc
import org.apache.spark.SparkException
/**
* An address identifier for an RPC endpoint.
*
* The `rpcAddress` may be null, in which case the endpoint is registered via a client-only
* connection and can only be... |
CrazyTechnology/spark | sql/catalyst/src/main/scala/org/apache/spark/sql/internal/ReadOnlySQLConf.scala | package org.apache.spark.sql.internal
import java.util.{Map => JMap}
import org.apache.spark.TaskContext
import org.apache.spark.internal.config.{ConfigEntry, ConfigProvider, ConfigReader}
/**
* A readonly SQLConf that will be created by tasks running at the executor side. It reads the
* configs from the local pro... |
CrazyTechnology/spark | core/src/test/scala/org/apache/spark/scheduler/ExternalClusterManagerSuite.scala |
package org.apache.spark.scheduler
import org.apache.spark.{LocalSparkContext, SparkConf, SparkContext, SparkFunSuite}
import org.apache.spark.executor.ExecutorMetrics
import org.apache.spark.scheduler.SchedulingMode.SchedulingMode
import org.apache.spark.storage.BlockManagerId
import org.apache.spark.util.Accumulato... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/TaskState.scala | <reponame>CrazyTechnology/spark
package org.apache.spark
private[spark] object TaskState extends Enumeration {
val LAUNCHING, RUNNING, FINISHED, FAILED, KILLED, LOST = Value
private val FINISHED_STATES = Set(FINISHED, FAILED, KILLED, LOST)
type TaskState = Value
def isFailed(state: TaskState): Boolean = (... |
CrazyTechnology/spark | sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/plans/logical/Command.scala | <gh_stars>0
package org.apache.spark.sql.catalyst.plans.logical
import org.apache.spark.sql.catalyst.expressions.Attribute
/**
* A logical node that represents a non-query command to be executed by the system. For example,
* commands can be used by parsers to represent DDL operations. Commands, unlike queries, a... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/deploy/master/DriverState.scala | <reponame>CrazyTechnology/spark
package org.apache.spark.deploy.master
private[deploy] object DriverState extends Enumeration {
type DriverState = Value
// SUBMITTED: Submitted but not yet scheduled on a worker
// RUNNING: Has been allocated to a worker to run
// FINISHED: Previously ran and exited cleanly
... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/deploy/history/FsHistoryProvider.scala |
package org.apache.spark.deploy.history
import java.io.{File, FileNotFoundException, IOException}
import java.nio.file.Files
import java.util.{Date, ServiceLoader}
import java.util.concurrent.{ConcurrentHashMap, ExecutorService, Future, TimeUnit}
import java.util.zip.{ZipEntry, ZipOutputStream}
import scala.collecti... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/Aggregator.scala | <gh_stars>0
package org.apache.spark
import org.apache.spark.annotation.DeveloperApi
import org.apache.spark.util.collection.ExternalAppendOnlyMap
/**
* :: DeveloperApi ::
* A set of functions used to aggregate data.
*
* @param createCombiner function to create the initial value of the aggregation.
* @param merg... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/deploy/master/ZooKeeperLeaderElectionAgent.scala |
package org.apache.spark.deploy.master
import org.apache.curator.framework.CuratorFramework
import org.apache.curator.framework.recipes.leader.{LeaderLatch, LeaderLatchListener}
import org.apache.spark.SparkConf
import org.apache.spark.deploy.SparkCuratorUtil
import org.apache.spark.internal.Logging
private[master]... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/deploy/master/WorkerState.scala | <filename>core/src/main/scala/org/apache/spark/deploy/master/WorkerState.scala<gh_stars>0
package org.apache.spark.deploy.master
private[master] object WorkerState extends Enumeration {
type WorkerState = Value
val ALIVE, DEAD, DECOMMISSIONED, UNKNOWN = Value
}
|
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/scheduler/DAGSchedulerSource.scala | <filename>core/src/main/scala/org/apache/spark/scheduler/DAGSchedulerSource.scala
package org.apache.spark.scheduler
import com.codahale.metrics.{Gauge, MetricRegistry, Timer}
import org.apache.spark.metrics.source.Source
private[scheduler] class DAGSchedulerSource(val dagScheduler: DAGScheduler)
extends Source ... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/util/taskListeners.scala |
package org.apache.spark.util
import java.util.EventListener
import org.apache.spark.TaskContext
import org.apache.spark.annotation.DeveloperApi
/**
* :: DeveloperApi ::
*
* Listener providing a callback function to invoke when a task's execution completes.
*/
@DeveloperApi
trait TaskCompletionListener extends ... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/deploy/master/ZooKeeperPersistenceEngine.scala | <gh_stars>0
package org.apache.spark.deploy.master
import java.nio.ByteBuffer
import scala.collection.JavaConverters._
import scala.reflect.ClassTag
import org.apache.curator.framework.CuratorFramework
import org.apache.zookeeper.CreateMode
import org.apache.spark.SparkConf
import org.apache.spark.deploy.SparkCura... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/scheduler/SchedulerBackend.scala | <reponame>CrazyTechnology/spark<filename>core/src/main/scala/org/apache/spark/scheduler/SchedulerBackend.scala
package org.apache.spark.scheduler
/**
* A backend interface for scheduling systems that allows plugging in different ones under
* TaskSchedulerImpl. We assume a Mesos-like model where the application gets ... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/broadcast/BroadcastManager.scala | <reponame>CrazyTechnology/spark
package org.apache.spark.broadcast
import java.util.concurrent.atomic.AtomicLong
import scala.reflect.ClassTag
import org.apache.commons.collections.map.{AbstractReferenceMap, ReferenceMap}
import org.apache.spark.{SecurityManager, SparkConf}
import org.apache.spark.internal.Logging... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/deploy/master/ApplicationState.scala |
package org.apache.spark.deploy.master
private[master] object ApplicationState extends Enumeration {
type ApplicationState = Value
val WAITING, RUNNING, FINISHED, FAILED, KILLED, UNKNOWN = Value
}
|
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/metrics/source/Source.scala | <reponame>CrazyTechnology/spark
package org.apache.spark.metrics.source
import com.codahale.metrics.MetricRegistry
private[spark] trait Source {
def sourceName: String
def metricRegistry: MetricRegistry
}
|
CrazyTechnology/spark | sql/hive/src/main/scala/org/apache/spark/sql/hive/package.scala | <gh_stars>0
package org.apache.spark.sql
/**
* Support for running Spark SQL queries using functionality from Apache Hive (does not require an
* existing Hive installation). Supported Hive features include:
* - Using HiveQL to express queries.
* - Reading metadata from the Hive Metastore using HiveSerDes.
* ... |
CrazyTechnology/spark | sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DataSourceV2Relation.scala | <reponame>CrazyTechnology/spark
package org.apache.spark.sql.execution.datasources.v2
import java.util.UUID
import scala.collection.JavaConverters._
import org.apache.spark.sql.{AnalysisException, SaveMode}
import org.apache.spark.sql.catalyst.analysis.{MultiInstanceRelation, NamedRelation}
import org.apache.spark.... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/deploy/client/StandaloneAppClientListener.scala |
package org.apache.spark.deploy.client
/**
* Callbacks invoked by deploy client when various events happen. There are currently five events:
* connecting to the cluster, disconnecting, being given an executor, having an executor removed
* (either due to failure or due to revocation), and having a worker removed.
... |
CrazyTechnology/spark | core/src/main/scala/org/apache/spark/scheduler/cluster/ExecutorData.scala | <gh_stars>0
package org.apache.spark.scheduler.cluster
import org.apache.spark.rpc.{RpcAddress, RpcEndpointRef}
/**
* Grouping of data for an executor used by CoarseGrainedSchedulerBackend.
*
* @param executorEndpoint The RpcEndpointRef representing this executor
* @param executorAddress The network address of t... |
lamastex/spark-trend-calculus-examples | notebooks/db/03streamable-trend-calculus-estimators.scala | // Databricks notebook source
// MAGIC %md
// MAGIC # Markov Model for Trend Calculus
// MAGIC
// MAGIC <NAME>, <NAME> and <NAME>
// MAGIC
// MAGIC 2020, Uppsala, Sweden
// MAGIC
// MAGIC This project was supported by Combient Mix AB through summer internships at:
// MAGIC
// MAGIC Combient Competence Centre for Da... |
lamastex/spark-trend-calculus-examples | notebooks/db/gdelt-POI-detection.scala | <reponame>lamastex/spark-trend-calculus-examples
// Databricks notebook source
// MAGIC %md
// MAGIC # Detecting Persons of Interest to OIL/GAS Price Trends
// MAGIC
// MAGIC <NAME>, <NAME> and <NAME>
// MAGIC
// MAGIC 2020, Uppsala, Sweden
// MAGIC
// MAGIC
// MAGIC This project was supported by Combient Mix AB th... |
lamastex/spark-trend-calculus-examples | notebooks/db/FX1M.scala | // Databricks notebook source
// MAGIC %md
// MAGIC # Historical Yahoo! Finance data
// MAGIC
// MAGIC <NAME>, <NAME> and <NAME>
// MAGIC
// MAGIC 2020, Uppsala, Sweden
// MAGIC
// MAGIC This project was supported by Combient Mix AB through summer internships at:
// MAGIC
// MAGIC Combient Competence Centre for Dat... |
lamastex/spark-trend-calculus-examples | notebooks/db/Overview.scala | <reponame>lamastex/spark-trend-calculus-examples
// Databricks notebook source
// MAGIC %md
// MAGIC # Trend Calculus of OIL Price
// MAGIC
// MAGIC <NAME>, <NAME> and <NAME>
// MAGIC
// MAGIC 2020, Uppsala, Sweden
// MAGIC
// MAGIC This work was inspired by <NAME>'s texata-2017 repository forked here:
// MAGIC
// ... |
lamastex/spark-trend-calculus-examples | notebooks/db/spark-gdelt-examples/gdelt-EOI-detection.scala | <reponame>lamastex/spark-trend-calculus-examples<filename>notebooks/db/spark-gdelt-examples/gdelt-EOI-detection.scala
// Databricks notebook source
// MAGIC %md
// MAGIC # Detecting Events of Interest to OIL/GAS Price Trends
// MAGIC
// MAGIC <NAME>, <NAME> and <NAME>
// MAGIC
// MAGIC 2020, Uppsala, Sweden
// MAGIC ... |
atomicbits/scramlgen | modules/scraml-raml-parser/src/main/scala/io/atomicbits/scraml/ramlparser/parser/KeyedList.scala | /*
*
* (C) Copyright 2018 Atomic BITS (http://atomicbits.io).
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by appl... |
atomicbits/scramlgen | modules/scraml-dsl-scala/src/main/scala/io/atomicbits/scraml/dsl/scalaplay/client/ClientConfig.scala | /*
*
* (C) Copyright 2018 Atomic BITS (http://atomicbits.io).
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by appl... |
atomicbits/scramlgen | modules/scraml-raml-parser/src/main/scala/io/atomicbits/scraml/ramlparser/model/parsedtypes/ParsedGenericObject.scala | <reponame>atomicbits/scramlgen
/*
*
* (C) Copyright 2018 Atomic BITS (http://atomicbits.io).
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
* http://www.apache.org/licenses/LICENSE-2.... |
atomicbits/scramlgen | modules/scraml-raml-parser/src/test/scala/io/atomicbits/scraml/ramlparser/model/FetchReplaceStringsTest.scala | <filename>modules/scraml-raml-parser/src/test/scala/io/atomicbits/scraml/ramlparser/model/FetchReplaceStringsTest.scala
/*
*
* (C) Copyright 2018 Atomic BITS (http://atomicbits.io).
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License... |
atomicbits/scramlgen | modules/scraml-generator/src/main/scala/io/atomicbits/scraml/generator/platform/javajackson/InterfaceGenerator.scala | /*
*
* (C) Copyright 2018 Atomic BITS (http://atomicbits.io).
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by appl... |
atomicbits/scramlgen | modules/scraml-raml-parser/src/main/scala/io/atomicbits/scraml/ramlparser/lookup/transformers/FallbackTransformer.scala | /*
*
* (C) Copyright 2018 Atomic BITS (http://atomicbits.io).
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by appl... |
atomicbits/scramlgen | modules/scraml-generator/src/test/scala/io/atomicbits/scraml/generator/WithEnumGeneratorTest.scala | /*
*
* (C) Copyright 2018 Atomic BITS (http://atomicbits.io).
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by appl... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.