object Pool extends Serializable
Companion object for Pool class that is CatBoost's abstraction of a dataset
- Alphabetic
- By Inheritance
- Pool
- Serializable
- Serializable
- AnyRef
- Any
- Hide All
- Show All
- Public
- All
Value Members
-
final
def
!=(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
final
def
##(): Int
- Definition Classes
- AnyRef → Any
-
final
def
==(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
final
def
asInstanceOf[T0]: T0
- Definition Classes
- Any
-
def
clone(): AnyRef
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws( ... ) @native()
-
final
def
eq(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
-
def
equals(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
def
finalize(): Unit
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws( classOf[java.lang.Throwable] )
-
final
def
getClass(): Class[_]
- Definition Classes
- AnyRef → Any
- Annotations
- @native()
-
def
hashCode(): Int
- Definition Classes
- AnyRef → Any
- Annotations
- @native()
-
final
def
isInstanceOf[T0]: Boolean
- Definition Classes
- Any
-
def
load(spark: SparkSession, dataPathWithScheme: String, columnDescription: Path = null, params: PoolLoadParams = new PoolLoadParams(), pairsDataPathWithScheme: String = null): Pool
Load dataset in one of CatBoost's natively supported formats:
- spark
SparkSession
- dataPathWithScheme
Path with scheme to dataset in CatBoost format. For example,
dsv:///home/user/datasets/my_dataset/train.dsv
orlibsvm:///home/user/datasets/my_dataset/train.libsvm
- columnDescription
Path to column description file
- params
Additional params specifying data format.
- pairsDataPathWithScheme
(optional) Path with scheme to dataset pairs in CatBoost format. Only "dsv-grouped" format is supported for now. For example,
dsv-grouped:///home/user/datasets/my_dataset/train_pairs.dsv
- returns
Pool containing loaded data.
val spark = SparkSession.builder() .master("local[*]") .appName("testLoadDSVSimple") .getOrCreate() val pool = Pool.load( spark, "dsv:///home/user/datasets/my_dataset/train.dsv", columnDescription = "/home/user/datasets/my_dataset/cd" ) val poolWithPairs = Pool.load( spark, "dsv:///home/user/datasets/my_dataset_with_pairs/train.dsv", columnDescription = "/home/user/datasets/my_dataset_with_pairs/cd", pairsDataPathWithScheme = "dsv-grouped:///home/user/datasets/my_dataset_with_pairs/train_pairs.dsv" )
Example: -
final
def
ne(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
-
final
def
notify(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native()
-
final
def
notifyAll(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native()
-
final
def
synchronized[T0](arg0: ⇒ T0): T0
- Definition Classes
- AnyRef
-
def
toString(): String
- Definition Classes
- AnyRef → Any
-
final
def
wait(): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... )
-
final
def
wait(arg0: Long, arg1: Int): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... )
-
final
def
wait(arg0: Long): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... ) @native()