Skip to content
Merged
Show file tree
Hide file tree
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Prev Previous commit
Centralize logic in a common AbstractSparkPlanUtils class
  • Loading branch information
charlesmyu committed Nov 6, 2025
commit 0bfaa064cdc78d1d0b372f6c94246d83efd8fc26
Original file line number Diff line number Diff line change
Expand Up @@ -24,6 +24,7 @@ public String[] helperClassNames() {
return new String[] {
packageName + ".AbstractDatadogSparkListener",
packageName + ".AbstractSparkPlanSerializer",
packageName + ".AbstractSparkPlanUtils",
packageName + ".DatabricksParentContext",
packageName + ".OpenlineageParentContext",
packageName + ".DatadogSpark212Listener",
Expand Down Expand Up @@ -109,7 +110,8 @@ public static void exit(
JavaConverters.mapAsScalaMap(planSerializer.extractFormattedProduct(plan))
.toMap(Predef.$conforms());

SparkPlanInfo newPlanInfo = Spark212PlanUtils.upsertSparkPlanInfoMetadata(planInfo, meta);
SparkPlanInfo newPlanInfo =
new Spark212PlanUtils().upsertSparkPlanInfoMetadata(planInfo, meta);
if (newPlanInfo != null) {
planInfo = newPlanInfo;
}
Expand Down
Original file line number Diff line number Diff line change
@@ -1,20 +1,11 @@
package datadog.trace.instrumentation.spark;

import datadog.trace.util.MethodHandles;
import java.lang.invoke.MethodHandle;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.List;
import org.apache.spark.sql.execution.SparkPlanInfo;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import scala.Option;
import scala.collection.immutable.Map;

public class Spark212PlanUtils {
private static final Logger log = LoggerFactory.getLogger(Spark212PlanUtils.class);

private static final MethodHandles methodLoader =
new MethodHandles(ClassLoader.getSystemClassLoader());
public class Spark212PlanUtils extends AbstractSparkPlanUtils {
private static final MethodHandle constructor =
methodLoader.constructor(
SparkPlanInfo.class,
Expand All @@ -35,43 +26,20 @@ public class Spark212PlanUtils {
String.class,
Option.class);

public static SparkPlanInfo upsertSparkPlanInfoMetadata(
SparkPlanInfo planInfo, scala.collection.immutable.Map<String, String> meta) {
// Attempt to create a new SparkPlanInfo with additional metadata replaced
// Since the fields are immutable we must instantiate a new SparkPlanInfo to do this

Object[] standardArgs =
new Object[] {
planInfo.nodeName(),
planInfo.simpleString(),
planInfo.children(),
meta,
planInfo.metrics()
};

if (databricksConstructor != null) {
List<Object> databricksArgs = new ArrayList<>(Arrays.asList(standardArgs));
try {
databricksArgs.add(SparkPlanInfo.class.getMethod("estRowCount").invoke(planInfo));
databricksArgs.add(SparkPlanInfo.class.getMethod("rddScopeId").invoke(planInfo));
databricksArgs.add(SparkPlanInfo.class.getMethod("explainId").invoke(planInfo));
} catch (Throwable t) {
log.warn("Error obtaining Databricks-specific SparkPlanInfo args", t);
}

SparkPlanInfo newPlan = methodLoader.invoke(databricksConstructor, databricksArgs.toArray());
if (newPlan != null) {
return newPlan;
}
}
@Override
protected MethodHandle getConstructor() {
return constructor;
}

if (constructor != null) {
SparkPlanInfo newPlan = methodLoader.invoke(constructor, standardArgs);
if (newPlan != null) {
return newPlan;
}
}
@Override
protected MethodHandle getDatabricksConstructor() {
return databricksConstructor;
}

return null;
@Override
protected Object[] getStandardArgs(SparkPlanInfo planInfo, Map meta) {
return new Object[] {
planInfo.nodeName(), planInfo.simpleString(), planInfo.children(), meta, planInfo.metrics()
};
}
}
Original file line number Diff line number Diff line change
Expand Up @@ -24,6 +24,7 @@ public String[] helperClassNames() {
return new String[] {
packageName + ".AbstractDatadogSparkListener",
packageName + ".AbstractSparkPlanSerializer",
packageName + ".AbstractSparkPlanUtils",
packageName + ".DatabricksParentContext",
packageName + ".OpenlineageParentContext",
packageName + ".DatadogSpark213Listener",
Expand Down Expand Up @@ -109,7 +110,8 @@ public static void exit(
Map<String, String> meta =
HashMap.from(JavaConverters.asScala(planSerializer.extractFormattedProduct(plan)));

SparkPlanInfo newPlanInfo = Spark213PlanUtils.upsertSparkPlanInfoMetadata(planInfo, meta);
SparkPlanInfo newPlanInfo =
new Spark213PlanUtils().upsertSparkPlanInfoMetadata(planInfo, meta);
if (newPlanInfo != null) {
planInfo = newPlanInfo;
}
Expand Down
Original file line number Diff line number Diff line change
@@ -1,20 +1,11 @@
package datadog.trace.instrumentation.spark;

import datadog.trace.util.MethodHandles;
import java.lang.invoke.MethodHandle;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.List;
import org.apache.spark.sql.execution.SparkPlanInfo;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import scala.Option;
import scala.collection.immutable.Map;

public class Spark213PlanUtils {
private static final Logger log = LoggerFactory.getLogger(Spark213PlanUtils.class);

private static final MethodHandles methodLoader =
new MethodHandles(ClassLoader.getSystemClassLoader());
public class Spark213PlanUtils extends AbstractSparkPlanUtils {
private static final MethodHandle constructor =
methodLoader.constructor(
SparkPlanInfo.class,
Expand All @@ -35,36 +26,20 @@ public class Spark213PlanUtils {
String.class,
Option.class);

public static SparkPlanInfo upsertSparkPlanInfoMetadata(
SparkPlanInfo planInfo, scala.collection.immutable.Map<String, String> meta) {
// Attempt to create a new SparkPlanInfo with additional metadata replaced
// Since the fields are immutable we must instantiate a new SparkPlanInfo to do this

Object[] standardArgs =
new Object[] {
planInfo.nodeName(),
planInfo.simpleString(),
planInfo.children(),
meta,
planInfo.metrics()
};

if (databricksConstructor != null) {
List<Object> databricksArgs = new ArrayList<>(Arrays.asList(standardArgs));
try {
databricksArgs.add(SparkPlanInfo.class.getMethod("estRowCount").invoke(planInfo));
databricksArgs.add(SparkPlanInfo.class.getMethod("rddScopeId").invoke(planInfo));
databricksArgs.add(SparkPlanInfo.class.getMethod("explainId").invoke(planInfo));
} catch (Throwable t) {
log.warn("Error obtaining Databricks-specific SparkPlanInfo args", t);
}
@Override
protected MethodHandle getConstructor() {
return constructor;
}

SparkPlanInfo newPlan = methodLoader.invoke(databricksConstructor, databricksArgs.toArray());
if (newPlan != null) {
return newPlan;
}
}
@Override
protected MethodHandle getDatabricksConstructor() {
return databricksConstructor;
}

return null;
@Override
protected Object[] getStandardArgs(SparkPlanInfo planInfo, Map meta) {
return new Object[] {
planInfo.nodeName(), planInfo.simpleString(), planInfo.children(), meta, planInfo.metrics()
};
}
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,57 @@
package datadog.trace.instrumentation.spark;

import datadog.trace.util.MethodHandles;
import java.lang.invoke.MethodHandle;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.List;
import org.apache.spark.sql.execution.SparkPlanInfo;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import scala.collection.immutable.Map;

abstract class AbstractSparkPlanUtils {
private static final Logger log = LoggerFactory.getLogger(AbstractSparkPlanUtils.class);

protected static final MethodHandles methodLoader =
new MethodHandles(ClassLoader.getSystemClassLoader());

protected abstract MethodHandle getConstructor();

protected abstract MethodHandle getDatabricksConstructor();

// Deals with Seq which changed from Scala 2.12 to 2.13, so delegate to version-specific classes
protected abstract Object[] getStandardArgs(SparkPlanInfo planInfo, Map meta);

// Attempt to create a new SparkPlanInfo with additional metadata replaced
// Since the fields are immutable we must instantiate a new SparkPlanInfo to do this
public SparkPlanInfo upsertSparkPlanInfoMetadata(
SparkPlanInfo planInfo, scala.collection.immutable.Map<String, String> meta) {
if (getDatabricksConstructor() != null) {
List<Object> databricksArgs = new ArrayList<>(Arrays.asList(getStandardArgs(planInfo, meta)));
try {
databricksArgs.add(SparkPlanInfo.class.getMethod("estRowCount").invoke(planInfo));
databricksArgs.add(SparkPlanInfo.class.getMethod("rddScopeId").invoke(planInfo));
databricksArgs.add(SparkPlanInfo.class.getMethod("explainId").invoke(planInfo));
} catch (Throwable t) {
log.warn("Error obtaining Databricks-specific SparkPlanInfo args", t);
}

SparkPlanInfo newPlan =
methodLoader.invoke(getDatabricksConstructor(), databricksArgs.toArray());
if (newPlan != null) {
return newPlan;
}
}

if (getConstructor() != null) {
SparkPlanInfo newPlan =
methodLoader.invoke(getConstructor(), getStandardArgs(planInfo, meta));
if (newPlan != null) {
return newPlan;
}
}

return null;
}
}
Loading