object Pool extends Serializable
Companion object for Pool class that is CatBoost's abstraction of a dataset
- Alphabetic
- By Inheritance
- Pool
- Serializable
- AnyRef
- Any
- Hide All
- Show All
- Public
- Protected
Value Members
- final def !=(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
- final def ##: Int
- Definition Classes
- AnyRef → Any
- final def ==(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
- final def asInstanceOf[T0]: T0
- Definition Classes
- Any
- def clone(): AnyRef
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws(classOf[java.lang.CloneNotSupportedException]) @native()
- final def eq(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
- def equals(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef → Any
- def finalize(): Unit
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws(classOf[java.lang.Throwable])
- final def getClass(): Class[_ <: AnyRef]
- Definition Classes
- AnyRef → Any
- Annotations
- @native()
- def hashCode(): Int
- Definition Classes
- AnyRef → Any
- Annotations
- @native()
- final def isInstanceOf[T0]: Boolean
- Definition Classes
- Any
- def load(spark: SparkSession, dataPathWithScheme: String, columnDescription: Path = null, params: PoolLoadParams = new PoolLoadParams(), pairsDataPathWithScheme: String = null): Pool
Load dataset in one of CatBoost's natively supported formats:
- spark
SparkSession
- dataPathWithScheme
Path with scheme to dataset in CatBoost format. For example,
dsv:///home/user/datasets/my_dataset/train.dsvorlibsvm:///home/user/datasets/my_dataset/train.libsvm- columnDescription
Path to column description file
- params
Additional params specifying data format.
- pairsDataPathWithScheme
(optional) Path with scheme to dataset pairs in CatBoost format. Only "dsv-grouped" format is supported for now. For example,
dsv-grouped:///home/user/datasets/my_dataset/train_pairs.dsv- returns
Pool containing loaded data.
val spark = SparkSession.builder() .master("local[*]") .appName("testLoadDSVSimple") .getOrCreate() val pool = Pool.load( spark, "dsv:///home/user/datasets/my_dataset/train.dsv", columnDescription = "/home/user/datasets/my_dataset/cd" ) val poolWithPairs = Pool.load( spark, "dsv:///home/user/datasets/my_dataset_with_pairs/train.dsv", columnDescription = "/home/user/datasets/my_dataset_with_pairs/cd", pairsDataPathWithScheme = "dsv-grouped:///home/user/datasets/my_dataset_with_pairs/train_pairs.dsv" )
Example: - final def ne(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
- final def notify(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native()
- final def notifyAll(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native()
- def read(spark: SparkSession): PoolReader
Returns a PoolReader that can be used to read Pool (API similar to Spark's DataFrameReader).
- final def synchronized[T0](arg0: => T0): T0
- Definition Classes
- AnyRef
- def toString(): String
- Definition Classes
- AnyRef → Any
- final def wait(): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws(classOf[java.lang.InterruptedException])
- final def wait(arg0: Long, arg1: Int): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws(classOf[java.lang.InterruptedException])
- final def wait(arg0: Long): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws(classOf[java.lang.InterruptedException]) @native()