export JAVA_TOOL_OPTIONS="-Djava.security.manager=allow -Dhadoop.security.token.service.use_ip=false" && \ export SPARK_LOCAL_IP=********* && \ export HADOOP_USER_NAME=$(whoami) && \ spark-shell --packages org.apache.arrow:arrow-memory:2.0.0,org.apache.arrow:arrow-vector:2.0.0 \ --conf "spark.hadoop.fs.defaultFS=file:///" import org.apache.spark.sql.SparkSession import org.apache.spark.sql.functions.{udf, lit} import org.apache.spark.sql.expressions.UserDefinedFunction object MandelbrotComparison { def main(args: Array[String]): Unit = { val spark = SparkSession.builder .appName("MandelbrotComparison") .getOrCreate() import spark.implicits._ // Configuración para optimizar el uso de memoria y partición spark.conf.set("spark.sql.inMemoryColumnarStorage.batchSize", "10000")…
Mandelbrot, Spark y UDF