t

scala.collection.parallel

ParIterableLike

trait ParIterableLike[+T, +CC[X] <: ParIterable[X], +Repr <: ParIterable[T], +Sequential <: scala.Iterable[T] with IterableOps[T, scala.Iterable, Sequential]] extends scala.IterableOnce[T] with CustomParallelizable[T, Repr] with Parallel with HasNewCombiner[T, Repr]

A template trait for parallel collections of type ParIterable[T].

This is a base trait for Scala parallel collections. It defines behaviour common to all parallel collections. Concrete parallel collections should inherit this trait and ParIterable if they want to define specific combiner factories.

Parallel operations are implemented with divide and conquer style algorithms that parallelize well. The basic idea is to split the collection into smaller parts until they are small enough to be operated on sequentially.

All of the parallel operations are implemented as tasks within this trait. Tasks rely on the concept of splitters, which extend iterators. Every parallel collection defines:

def splitter: IterableSplitter[T]

which returns an instance of IterableSplitter[T], which is a subtype of Splitter[T]. Splitters have a method remaining to check the remaining number of elements, and method split which is defined by splitters. Method split divides the splitters iterate over into disjunct subsets:

def split: Seq[Splitter]

which splits the splitter into a sequence of disjunct subsplitters. This is typically a very fast operation which simply creates wrappers around the receiver collection. This can be repeated recursively.

Tasks are scheduled for execution through a scala.collection.parallel.TaskSupport object, which can be changed through the tasksupport setter of the collection.

Method newCombiner produces a new combiner. Combiners are an extension of builders. They provide a method combine which combines two combiners and returns a combiner containing elements of both combiners. This method can be implemented by aggressively copying all the elements into the new combiner or by lazily binding their results. It is recommended to avoid copying all of the elements for performance reasons, although that cost might be negligible depending on the use case. Standard parallel collection combiners avoid copying when merging results, relying either on a two-step lazy construction or specific data-structure properties.

Methods:

def seq: Sequential
def par: Repr

produce the sequential or parallel implementation of the collection, respectively. Method par just returns a reference to this parallel collection. Method seq is efficient - it will not copy the elements. Instead, it will create a sequential version of the collection using the same underlying data structure. Note that this is not the case for sequential collections in general - they may copy the elements and produce a different underlying data structure.

The combination of methods toMap, toSeq or toSet along with par and seq is a flexible way to change between different collection types.

Since this trait extends the GenIterable trait, methods like size must also be implemented in concrete collections, while iterator forwards to splitter by default.

Each parallel collection is bound to a specific fork/join pool, on which dormant worker threads are kept. The fork/join pool contains other information such as the parallelism level, that is, the number of processors used. When a collection is created, it is assigned the default fork/join pool found in the scala.parallel package object.

Parallel collections are not necessarily ordered in terms of the foreach operation (see Traversable). Parallel sequences have a well defined order for iterators - creating an iterator and traversing the elements linearly will always yield the same order. However, bulk operations such as foreach, map or filter always occur in undefined orders for all parallel collections.

Existing parallel collection implementations provide strict parallel iterators. Strict parallel iterators are aware of the number of elements they have yet to traverse. It's also possible to provide non-strict parallel iterators, which do not know the number of elements remaining. To do this, the new collection implementation must override isStrictSplitterCollection to false. This will make some operations unavailable.

To create a new parallel collection, extend the ParIterable trait, and implement size, splitter, newCombiner and seq. Having an implicit combiner factory requires extending this trait in addition, as well as providing a companion object, as with regular collections.

Method size is implemented as a constant time operation for parallel collections, and parallel collection operations rely on this assumption.

The higher-order functions passed to certain operations may contain side-effects. Since implementations of bulk operations may not be sequential, this means that side-effects may not be predictable and may produce data-races, deadlocks or invalidation of state if care is not taken. It is up to the programmer to either avoid using side-effects or to use some form of synchronization when accessing mutable data.

T

the element type of the collection

Repr

the type of the actual collection containing the elements

Self Type
ParIterableLike[T, CC, Repr, Sequential]
Linear Supertypes
HasNewCombiner[T @scala.annotation.unchecked.uncheckedVariance, Repr], Parallel, CustomParallelizable[T @scala.annotation.unchecked.uncheckedVariance, Repr], Parallelizable[T @scala.annotation.unchecked.uncheckedVariance, Repr], IterableOnce[T @scala.annotation.unchecked.uncheckedVariance], AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. ParIterableLike
  2. HasNewCombiner
  3. Parallel
  4. CustomParallelizable
  5. Parallelizable
  6. IterableOnce
  7. AnyRef
  8. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. Protected

Type Members

  1. trait Accessor[R, Tp] extends StrictSplitterCheckTask[R, Tp]

    Standard accessor task that iterates over the elements of the collection.

    Standard accessor task that iterates over the elements of the collection.

    R

    type of the result of this method (R for result).

    Tp

    the representation type of the task at hand.

    Attributes
    protected
  2. class Aggregate[S] extends Accessor[S, Aggregate[S]]
    Attributes
    protected[this]
  3. trait BuilderOps[Elem, To] extends AnyRef
  4. class Collect[S, That] extends Transformer[Combiner[S, That], Collect[S, That]]
    Attributes
    protected[this]
  5. abstract class Composite[FR, SR, R, First <: StrictSplitterCheckTask[FR, _], Second <: StrictSplitterCheckTask[SR, _]] extends NonDivisibleTask[R, Composite[FR, SR, R, First, Second]]
    Attributes
    protected[this]
  6. class Copy[U >: T, That] extends Transformer[Combiner[U, That], Copy[U, That]]
    Attributes
    protected
  7. class CopyToArray[U >: T, This >: Repr] extends Accessor[Unit, CopyToArray[U, This]]
    Attributes
    protected[this]
  8. class Count extends Accessor[Int, Count]
    Attributes
    protected[this]
  9. class CreateScanTree[U >: T] extends Transformer[ScanTree[U], CreateScanTree[U]]
    Attributes
    protected[this]
  10. class Drop[U >: T, This >: Repr] extends Transformer[Combiner[U, This], Drop[U, This]]
    Attributes
    protected[this]
  11. class Exists extends Accessor[Boolean, Exists]
    Attributes
    protected[this]
  12. class Filter[U >: T, This >: Repr] extends Transformer[Combiner[U, This], Filter[U, This]]
    Attributes
    protected[this]
  13. class FilterNot[U >: T, This >: Repr] extends Transformer[Combiner[U, This], FilterNot[U, This]]
    Attributes
    protected[this]
  14. class Find[U >: T] extends Accessor[Option[U], Find[U]]
    Attributes
    protected[this]
  15. class FlatMap[S, That] extends Transformer[Combiner[S, That], FlatMap[S, That]]
    Attributes
    protected[this]
  16. class Fold[U >: T] extends Accessor[U, Fold[U]]
    Attributes
    protected[this]
  17. class Forall extends Accessor[Boolean, Forall]
    Attributes
    protected[this]
  18. class Foreach[S] extends Accessor[Unit, Foreach[S]]
    Attributes
    protected[this]
  19. class FromScanTree[U >: T, That] extends StrictSplitterCheckTask[Combiner[U, That], FromScanTree[U, That]]
    Attributes
    protected[this]
  20. class GroupBy[K, U >: T] extends Transformer[HashMapCombiner[K, U], GroupBy[K, U]]
    Attributes
    protected[this]
  21. class Map[S, That] extends Transformer[Combiner[S, That], Map[S, That]]
    Attributes
    protected[this]
  22. class Max[U >: T] extends Accessor[Option[U], Max[U]]
    Attributes
    protected[this]
  23. class Min[U >: T] extends Accessor[Option[U], Min[U]]
    Attributes
    protected[this]
  24. trait NonDivisible[R] extends NonDivisibleTask[R, NonDivisible[R]]
    Attributes
    protected[this]
  25. trait NonDivisibleTask[R, Tp] extends StrictSplitterCheckTask[R, Tp]
    Attributes
    protected[this]
  26. abstract class ParComposite[FR, SR, R, First <: StrictSplitterCheckTask[FR, _], Second <: StrictSplitterCheckTask[SR, _]] extends Composite[FR, SR, R, First, Second]

    Performs two tasks in parallel, and waits for both to finish.

    Performs two tasks in parallel, and waits for both to finish.

    Attributes
    protected[this]
  27. class Partition[U >: T, This >: Repr] extends Transformer[(Combiner[U, This], Combiner[U, This]), Partition[U, This]]
    Attributes
    protected[this]
  28. class Product[U >: T] extends Accessor[U, Product[U]]
    Attributes
    protected[this]
  29. class Reduce[U >: T] extends Accessor[Option[U], Reduce[U]]
    Attributes
    protected[this]
  30. abstract class ResultMapping[R, Tp, R1] extends NonDivisibleTask[R1, ResultMapping[R, Tp, R1]]
    Attributes
    protected[this]
  31. type SSCTask[R, Tp] = StrictSplitterCheckTask[R, Tp]
  32. case class ScanLeaf[U >: T](pit: IterableSplitter[U], op: (U, U) => U, from: Int, len: Int, prev: Option[ScanLeaf[U]], acc: U) extends ScanTree[U] with scala.Product with Serializable
    Attributes
    protected[this]
  33. case class ScanNode[U >: T](left: ScanTree[U], right: ScanTree[U]) extends ScanTree[U] with scala.Product with Serializable
    Attributes
    protected[this]
  34. sealed trait ScanTree[U >: T] extends AnyRef
    Attributes
    protected[this]
  35. abstract class SeqComposite[FR, SR, R, First <: StrictSplitterCheckTask[FR, _], Second <: StrictSplitterCheckTask[SR, _]] extends Composite[FR, SR, R, First, Second]

    Sequentially performs one task after another.

    Sequentially performs one task after another.

    Attributes
    protected[this]
  36. trait SignallingOps[PI <: DelegatedSignalling] extends AnyRef
  37. class Slice[U >: T, This >: Repr] extends Transformer[Combiner[U, This], Slice[U, This]]
    Attributes
    protected[this]
  38. class Span[U >: T, This >: Repr] extends Transformer[(Combiner[U, This], Combiner[U, This]), Span[U, This]]
    Attributes
    protected[this]
  39. class SplitAt[U >: T, This >: Repr] extends Transformer[(Combiner[U, This], Combiner[U, This]), SplitAt[U, This]]
    Attributes
    protected[this]
  40. trait StrictSplitterCheckTask[R, Tp] extends Task[R, Tp]
    Attributes
    protected
  41. class Sum[U >: T] extends Accessor[U, Sum[U]]
    Attributes
    protected[this]
  42. class Take[U >: T, This >: Repr] extends Transformer[Combiner[U, This], Take[U, This]]
    Attributes
    protected[this]
  43. class TakeWhile[U >: T, This >: Repr] extends Transformer[(Combiner[U, This], Boolean), TakeWhile[U, This]]
    Attributes
    protected[this]
  44. trait TaskOps[R, Tp] extends AnyRef
  45. class ToParCollection[U >: T, That] extends Transformer[Combiner[U, That], ToParCollection[U, That]]
    Attributes
    protected[this]
  46. class ToParMap[K, V, That] extends Transformer[Combiner[(K, V), That], ToParMap[K, V, That]]
    Attributes
    protected[this]
  47. trait Transformer[R, Tp] extends Accessor[R, Tp]
    Attributes
    protected
  48. class Zip[U >: T, S, That] extends Transformer[Combiner[(U, S), That], Zip[U, S, That]]
    Attributes
    protected[this]
  49. class ZipAll[U >: T, S, That] extends Transformer[Combiner[(U, S), That], ZipAll[U, S, That]]
    Attributes
    protected[this]

Abstract Value Members

  1. abstract def companion: GenericParCompanion[CC]
  2. abstract def newCombiner: Combiner[T @scala.annotation.unchecked.uncheckedVariance, Repr]
    Attributes
    protected[this]
    Definition Classes
    HasNewCombiner
  3. abstract def seq: Sequential
    Definition Classes
    ParIterableLikeParallelizable
  4. abstract def size: Int
  5. abstract def splitter: IterableSplitter[T]

    Creates a new parallel iterator used to traverse the elements of this parallel collection.

    Creates a new parallel iterator used to traverse the elements of this parallel collection. This iterator is more specific than the iterator of the returned by iterator, and augmented with additional accessor and transformer methods.

    returns

    a parallel iterator

    Attributes
    protected[parallel]
  6. abstract def stringPrefix: String

Concrete Value Members

  1. final def !=(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  2. final def ##: Int
    Definition Classes
    AnyRef → Any
  3. def ++[U >: T](that: scala.IterableOnce[U]): CC[U]
  4. def /:[S](z: S)(op: (S, T) => S): S
  5. def :\[S](z: S)(op: (T, S) => S): S
  6. final def ==(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  7. def aggregate[S](z: => S)(seqop: (S, T) => S, combop: (S, S) => S): S

    Aggregates the results of applying an operator to subsequent elements.

    Aggregates the results of applying an operator to subsequent elements.

    This is a more general form of fold and reduce. It has similar semantics, but does not require the result to be a supertype of the element type. It traverses the elements in different partitions sequentially, using seqop to update the result, and then applies combop to results from different partitions. The implementation of this operation may operate on an arbitrary number of collection partitions, so combop may be invoked arbitrary number of times.

    For example, one might want to process some elements and then produce a Set. In this case, seqop would process an element and append it to the set, while combop would concatenate two sets from different partitions together. The initial value z would be an empty set.

    pc.aggregate(Set[Int]())(_ += process(_), _ ++ _)

    Another example is calculating geometric mean from a collection of doubles (one would typically require big doubles for this).

    S

    the type of accumulated results

    z

    the initial value for the accumulated result of the partition - this will typically be the neutral element for the seqop operator (e.g. Nil for list concatenation or 0 for summation) and may be evaluated more than once

    seqop

    an operator used to accumulate results within a partition

    combop

    an associative operator used to combine results from different partitions

  8. final def asInstanceOf[T0]: T0
    Definition Classes
    Any
  9. implicit def builder2ops[Elem, To](cb: Builder[Elem, To]): BuilderOps[Elem, To]
    Attributes
    protected
  10. def clone(): AnyRef
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.CloneNotSupportedException]) @native()
  11. def collect[S](pf: PartialFunction[T, S]): CC[S]
  12. def combinerFactory[S, That](cbf: () => Combiner[S, That]): CombinerFactory[S, That]
    Attributes
    protected[this]
  13. def combinerFactory: CombinerFactory[T, Repr]

    Creates a combiner factory.

    Creates a combiner factory. Each combiner factory instance is used once per invocation of a parallel transformer method for a single collection.

    The default combiner factory creates a new combiner every time it is requested, unless the combiner is thread-safe as indicated by its canBeShared method. In this case, the method returns a factory which returns the same combiner each time. This is typically done for concurrent parallel collections, the combiners of which allow thread safe access.

    Attributes
    protected[this]
  14. def copyToArray[U >: T](xs: Array[U], start: Int, len: Int): Unit
  15. def copyToArray[U >: T](xs: Array[U], start: Int): Unit
  16. def copyToArray[U >: T](xs: Array[U]): Unit
  17. def count(p: (T) => Boolean): Int
  18. def debugBuffer: ArrayBuffer[String]
  19. implicit def delegatedSignalling2ops[PI <: DelegatedSignalling](it: PI): SignallingOps[PI]
    Attributes
    protected
  20. def drop(n: Int): Repr
  21. def dropWhile(pred: (T) => Boolean): Repr

    Drops all elements in the longest prefix of elements that satisfy the predicate, and returns a collection composed of the remaining elements.

    Drops all elements in the longest prefix of elements that satisfy the predicate, and returns a collection composed of the remaining elements.

    This method will use indexFlag signalling capabilities. This means that splitters may set and read the indexFlag state. The index flag is initially set to maximum integer value.

    pred

    the predicate used to test the elements

    returns

    a collection composed of all the elements after the longest prefix of elements in this parallel iterable that satisfy the predicate pred

  22. final def eq(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  23. def equals(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef → Any
  24. def exists(p: (T) => Boolean): Boolean

    Tests whether a predicate holds for some element of this parallel iterable.

    Tests whether a predicate holds for some element of this parallel iterable.

    This method will use abort signalling capabilities. This means that splitters may send and read abort signals.

    p

    a predicate used to test elements

    returns

    true if p holds for some element, false otherwise

  25. def filter(pred: (T) => Boolean): Repr
  26. def filterNot(pred: (T) => Boolean): Repr
  27. def finalize(): Unit
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.Throwable])
  28. def find(p: (T) => Boolean): Option[T]

    Finds some element in the collection for which the predicate holds, if such an element exists.

    Finds some element in the collection for which the predicate holds, if such an element exists. The element may not necessarily be the first such element in the iteration order.

    If there are multiple elements obeying the predicate, the choice is nondeterministic.

    This method will use abort signalling capabilities. This means that splitters may send and read abort signals.

    p

    predicate used to test the elements

    returns

    an option value with the element if such an element exists, or None otherwise

  29. def flatMap[S](f: (T) => scala.IterableOnce[S]): CC[S]
  30. def fold[U >: T](z: U)(op: (U, U) => U): U

    Folds the elements of this sequence using the specified associative binary operator.

    Folds the elements of this sequence using the specified associative binary operator. The order in which the elements are reduced is unspecified and may be nondeterministic.

    Note this method has a different signature than the foldLeft and foldRight methods of the trait Traversable. The result of folding may only be a supertype of this parallel collection's type parameter T.

    U

    a type parameter for the binary operator, a supertype of T.

    z

    a neutral element for the fold operation, it may be added to the result an arbitrary number of times, not changing the result (e.g. Nil for list concatenation, 0 for addition, or 1 for multiplication)

    op

    a binary operator that must be associative

    returns

    the result of applying fold operator op between all the elements and z

  31. def foldLeft[S](z: S)(op: (S, T) => S): S
  32. def foldRight[S](z: S)(op: (T, S) => S): S
  33. def forall(p: (T) => Boolean): Boolean

    Tests whether a predicate holds for all elements of this parallel iterable.

    Tests whether a predicate holds for all elements of this parallel iterable.

    This method will use abort signalling capabilities. This means that splitters may send and read abort signals.

    p

    a predicate used to test elements

    returns

    true if p holds for all elements, false otherwise

  34. def foreach[U](f: (T) => U): Unit

    Applies a function f to all the elements of parallel iterable in an undefined order.

    Applies a function f to all the elements of parallel iterable in an undefined order.

    U

    the result type of the function applied to each element, which is always discarded

    f

    function applied to each element

  35. final def getClass(): Class[_ <: AnyRef]
    Definition Classes
    AnyRef → Any
    Annotations
    @native()
  36. def groupBy[K](f: (T) => K): immutable.ParMap[K, Repr]
  37. def hasDefiniteSize: Boolean
  38. def hashCode(): Int
    Definition Classes
    AnyRef → Any
    Annotations
    @native()
  39. def head: T
  40. def headOption: Option[T]
  41. def init: Repr
  42. def initTaskSupport(): Unit
    Attributes
    protected
  43. def isEmpty: Boolean
  44. final def isInstanceOf[T0]: Boolean
    Definition Classes
    Any
  45. def isStrictSplitterCollection: Boolean

    Denotes whether this parallel collection has strict splitters.

    Denotes whether this parallel collection has strict splitters.

    This is true in general, and specific collection instances may choose to override this method. Such collections will fail to execute methods which rely on splitters being strict, i.e. returning a correct value in the remaining method.

    This method helps ensure that such failures occur on method invocations, rather than later on and in unpredictable ways.

  46. final def isTraversableAgain: Boolean
  47. def iterator: Splitter[T]

    Creates a new split iterator used to traverse the elements of this collection.

    Creates a new split iterator used to traverse the elements of this collection.

    By default, this method is implemented in terms of the protected splitter method.

    returns

    a split iterator

    Definition Classes
    ParIterableLike → IterableOnce
  48. def knownSize: Int
    Definition Classes
    IterableOnce
  49. def last: T
  50. def lastOption: Option[T]
  51. def map[S](f: (T) => S): CC[S]
  52. def max[U >: T](implicit ord: Ordering[U]): T
  53. def maxBy[S](f: (T) => S)(implicit cmp: Ordering[S]): T
  54. def min[U >: T](implicit ord: Ordering[U]): T
  55. def minBy[S](f: (T) => S)(implicit cmp: Ordering[S]): T
  56. def mkString: String
  57. def mkString(sep: String): String
  58. def mkString(start: String, sep: String, end: String): String
  59. final def ne(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  60. def nonEmpty: Boolean
  61. final def notify(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native()
  62. final def notifyAll(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native()
  63. def par: Repr

    Returns a parallel implementation of this collection.

    Returns a parallel implementation of this collection.

    For most collection types, this method creates a new parallel collection by copying all the elements. For these collection, par takes linear time. Mutable collections in this category do not produce a mutable parallel collection that has the same underlying dataset, so changes in one collection will not be reflected in the other one.

    Specific collections (e.g. ParArray or mutable.ParHashMap) override this default behaviour by creating a parallel collection which shares the same underlying dataset. For these collections, par takes constant or sublinear time.

    All parallel collections return a reference to themselves.

    returns

    a parallel implementation of this collection

    Definition Classes
    ParIterableLikeCustomParallelizableParallelizable
  64. def parCombiner: Nothing

    The default par implementation uses the combiner provided by this method to create a new parallel collection.

    The default par implementation uses the combiner provided by this method to create a new parallel collection.

    returns

    a combiner for the parallel collection of type ParRepr

    Attributes
    protected[this]
    Definition Classes
    CustomParallelizableParallelizable
  65. def partition(pred: (T) => Boolean): (Repr, Repr)
  66. def product[U >: T](implicit num: Numeric[U]): U
  67. def reduce[U >: T](op: (U, U) => U): U

    Reduces the elements of this sequence using the specified associative binary operator.

    Reduces the elements of this sequence using the specified associative binary operator.

    The order in which operations are performed on elements is unspecified and may be nondeterministic.

    Note this method has a different signature than the reduceLeft and reduceRight methods of the trait Traversable. The result of reducing may only be a supertype of this parallel collection's type parameter T.

    U

    A type parameter for the binary operator, a supertype of T.

    op

    A binary operator that must be associative.

    returns

    The result of applying reduce operator op between all the elements if the collection is nonempty.

    Exceptions thrown

    UnsupportedOperationException if this parallel iterable is empty.

  68. def reduceLeft[U >: T](op: (U, T) => U): U
  69. def reduceLeftOption[U >: T](op: (U, T) => U): Option[U]
  70. def reduceOption[U >: T](op: (U, U) => U): Option[U]

    Optionally reduces the elements of this sequence using the specified associative binary operator.

    Optionally reduces the elements of this sequence using the specified associative binary operator.

    The order in which operations are performed on elements is unspecified and may be nondeterministic.

    Note this method has a different signature than the reduceLeftOption and reduceRightOption methods of the trait Traversable. The result of reducing may only be a supertype of this parallel collection's type parameter T.

    U

    A type parameter for the binary operator, a supertype of T.

    op

    A binary operator that must be associative.

    returns

    An option value containing result of applying reduce operator op between all the elements if the collection is nonempty, and None otherwise.

  71. def reduceRight[U >: T](op: (T, U) => U): U
  72. def reduceRightOption[U >: T](op: (T, U) => U): Option[U]
  73. def repr: Repr
  74. def reuse[S, That](oldc: Option[Combiner[S, That]], newc: Combiner[S, That]): Combiner[S, That]

    Optionally reuses an existing combiner for better performance.

    Optionally reuses an existing combiner for better performance. By default it doesn't - subclasses may override this behaviour. The provided combiner oldc that can potentially be reused will be either some combiner from the previous computational task, or None if there was no previous phase (in which case this method must return newc).

    oldc

    The combiner that is the result of the previous task, or None if there was no previous task.

    newc

    The new, empty combiner that can be used.

    returns

    Either newc or oldc.

    Attributes
    protected
  75. def sameElements[U >: T](that: scala.IterableOnce[U]): Boolean
  76. def scan[U >: T](z: U)(op: (U, U) => U): CC[U]

    Computes a prefix scan of the elements of the collection.

    Computes a prefix scan of the elements of the collection.

    Note: The neutral element z may be applied more than once.

    U

    element type of the resulting collection

    z

    neutral element for the operator op

    op

    the associative operator for the scan

    returns

    a new parallel iterable containing the prefix scan of the elements in this parallel iterable

  77. def scanBlockSize: Int
    Attributes
    protected[this]
  78. def scanLeft[S](z: S)(op: (S, T) => S): scala.Iterable[S]
  79. def scanRight[S](z: S)(op: (T, S) => S): scala.Iterable[S]
  80. def sequentially[S, That <: Parallel](b: (Sequential) => Sequential): Repr
    Attributes
    protected[this]
  81. def slice(unc_from: Int, unc_until: Int): Repr
  82. def span(pred: (T) => Boolean): (Repr, Repr)

    Splits this parallel iterable into a prefix/suffix pair according to a predicate.

    Splits this parallel iterable into a prefix/suffix pair according to a predicate.

    This method will use indexFlag signalling capabilities. This means that splitters may set and read the indexFlag state. The index flag is initially set to maximum integer value.

    pred

    the predicate used to test the elements

    returns

    a pair consisting of the longest prefix of the collection for which all the elements satisfy pred, and the rest of the collection

  83. def splitAt(n: Int): (Repr, Repr)
  84. def stepper[S <: Stepper[_]](implicit shape: StepperShape[T @scala.annotation.unchecked.uncheckedVariance, S]): S
    Definition Classes
    IterableOnce
  85. def sum[U >: T](implicit num: Numeric[U]): U
  86. final def synchronized[T0](arg0: => T0): T0
    Definition Classes
    AnyRef
  87. def tail: Repr
  88. def take(n: Int): Repr
  89. def takeWhile(pred: (T) => Boolean): Repr

    Takes the longest prefix of elements that satisfy the predicate.

    Takes the longest prefix of elements that satisfy the predicate.

    This method will use indexFlag signalling capabilities. This means that splitters may set and read the indexFlag state. The index flag is initially set to maximum integer value.

    pred

    the predicate used to test the elements

    returns

    the longest prefix of this parallel iterable of elements that satisfy the predicate pred

  90. implicit def task2ops[R, Tp](tsk: SSCTask[R, Tp]): TaskOps[R, Tp]
    Attributes
    protected
  91. def tasksupport: TaskSupport

    The task support object which is responsible for scheduling and load-balancing tasks to processors.

    The task support object which is responsible for scheduling and load-balancing tasks to processors.

    See also

    scala.collection.parallel.TaskSupport

  92. def tasksupport_=(ts: TaskSupport): Unit

    Changes the task support object which is responsible for scheduling and load-balancing tasks to processors.

    Changes the task support object which is responsible for scheduling and load-balancing tasks to processors.

    A task support object can be changed in a parallel collection after it has been created, but only during a quiescent period, i.e. while there are no concurrent invocations to parallel collection methods.

    Here is a way to change the task support of a parallel collection:

    import scala.collection.parallel._
    val pc = mutable.ParArray(1, 2, 3)
    pc.tasksupport = new ForkJoinTaskSupport(
      new java.util.concurrent.ForkJoinPool(2))
    See also

    scala.collection.parallel.TaskSupport

  93. def to[C](factory: Factory[T, C]): C
  94. def toArray[U >: T](implicit arg0: ClassTag[U]): Array[U]
  95. def toBuffer[U >: T]: Buffer[U]
  96. def toIndexedSeq: immutable.IndexedSeq[T]
  97. def toIterable: ParIterable[T]
  98. def toIterator: scala.Iterator[T]
  99. def toList: List[T]
  100. def toMap[K, V](implicit ev: <:<[T, (K, V)]): immutable.ParMap[K, V]
  101. def toParCollection[U >: T, That](cbf: () => Combiner[U, That]): That
    Attributes
    protected
  102. def toParMap[K, V, That](cbf: () => Combiner[(K, V), That])(implicit ev: <:<[T, (K, V)]): That
    Attributes
    protected
  103. def toSeq: ParSeq[T]
  104. def toSet[U >: T]: immutable.ParSet[U]
  105. def toString(): String
    Definition Classes
    ParIterableLike → AnyRef → Any
  106. def toVector: Vector[T]
  107. final def wait(): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.InterruptedException])
  108. final def wait(arg0: Long, arg1: Int): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.InterruptedException])
  109. final def wait(arg0: Long): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.InterruptedException]) @native()
  110. def withFilter(pred: (T) => Boolean): Repr
  111. def wrap[R](body: => R): NonDivisible[R]
    Attributes
    protected
  112. def zip[U >: T, S](that: scala.Iterable[S]): CC[(U, S)]
  113. def zip[U >: T, S](that: ParIterable[S]): CC[(U, S)]
  114. def zipAll[S, U >: T](that: ParIterable[S], thisElem: U, thatElem: S): CC[(U, S)]
  115. def zipWithIndex[U >: T]: CC[(U, Int)]

    Zips this parallel iterable with its indices.

    Zips this parallel iterable with its indices.

    U

    the type of the first half of the returned pairs (this is always a supertype of the collection's element type T).

    returns

    A new collection of type ParIterable containing pairs consisting of all elements of this parallel iterable paired with their index. Indices start at 0.

Deprecated Value Members

  1. def toStream: Stream[T]
    Annotations
    @deprecated
    Deprecated

    (Since version 0.1.3) Use to(LazyList) instead.

  2. def toTraversable: ParIterable[T]
    Annotations
    @deprecated
    Deprecated

    (Since version 0.1.3) Use toIterable instead

Inherited from HasNewCombiner[T @scala.annotation.unchecked.uncheckedVariance, Repr]

Inherited from Parallel

Inherited from CustomParallelizable[T @scala.annotation.unchecked.uncheckedVariance, Repr]

Inherited from Parallelizable[T @scala.annotation.unchecked.uncheckedVariance, Repr]

Inherited from IterableOnce[T @scala.annotation.unchecked.uncheckedVariance]

Inherited from AnyRef

Inherited from Any

Ungrouped