Is there a way to implement dimension reduction using pyspark? I have a dataframe and loaded into pyspark.
FILENAME = "test.csv"
spark = SparkSession.builder.appName('Test') \
.getOrCreate()
spark_df = spark.read.csv(FILENAME, header=True)
# Load the embeddings from the spark_df
embedded_df_columns = spark_df.columns[5:]
embedded_df = spark_df.select(embedded_df_columns)
I dont seem to find the right pyspark.ml.features for tsne. All I get was for pca. Can anyone help please