@@ -84,7 +84,10 @@ def submit_feathr_job(
8484 cfg = configuration .copy () if configuration else {}
8585 maven_dependency = f"{ cfg .pop ('spark.jars.packages' , self .packages )} ,{ get_maven_artifact_fullname ()} "
8686 spark_args = self ._init_args (job_name = job_name , confs = cfg )
87-
87+ # Add additional repositories
88+ spark_args .extend (["--repositories" , "https://repository.mulesoft.org/nexus/content/repositories/public/,https://linkedin.jfrog.io/artifactory/open-source/" ])
89+ # spark_args.extend(["--repositories", "https://linkedin.jfrog.io/artifactory/open-source/"])
90+
8891 if not main_jar_path :
8992 # We don't have the main jar, use Maven
9093 if not python_files :
@@ -106,7 +109,16 @@ def submit_feathr_job(
106109 print (python_files )
107110 spark_args .append (python_files [0 ])
108111 else :
109- spark_args .extend (["--class" , main_class_name , main_jar_path ])
112+ if not python_files :
113+ # This is a JAR job
114+ spark_args .extend (["--class" , main_class_name , main_jar_path ])
115+ else :
116+ spark_args .extend (["--packages" , maven_dependency ])
117+ # This is a PySpark job, no more things to
118+ if python_files .__len__ () > 1 :
119+ spark_args .extend (["--py-files" , "," .join (python_files [1 :])])
120+ spark_args .append (python_files [0 ])
121+
110122
111123 if arguments :
112124 spark_args .extend (arguments )
@@ -299,4 +311,5 @@ def _get_default_package(self):
299311 packages .append ("commons-io:commons-io:2.6" )
300312 packages .append ("org.apache.hadoop:hadoop-azure:2.7.4" )
301313 packages .append ("com.microsoft.azure:azure-storage:8.6.4" )
314+ packages .append ("com.github.everit-org.json-schema:org.everit.json.schema:1.9.1" )
302315 return "," .join (packages )
0 commit comments