Skip to content

Commit de5f2bb

Browse files
authored
chore: remove unused import (#246)
1 parent d66ffe4 commit de5f2bb

30 files changed

+49
-130
lines changed

spark/.scalafix.conf

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,3 @@
1+
rules = [
2+
RemoveUnused
3+
]

spark/pom.xml

Lines changed: 23 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -197,11 +197,29 @@
197197
<groupId>net.alchim31.maven</groupId>
198198
<artifactId>scala-maven-plugin</artifactId>
199199
<version>4.8.0</version>
200+
<executions>
201+
<execution>
202+
<goals>
203+
<goal>compile</goal>
204+
<goal>testCompile</goal>
205+
</goals>
206+
</execution>
207+
</executions>
200208
<configuration>
201209
<jvmArgs>
202210
<jvmArg>-Xms64m</jvmArg>
203211
<jvmArg>-Xmx1024m</jvmArg>
204212
</jvmArgs>
213+
<args>
214+
<arg>-Ywarn-unused</arg>
215+
</args>
216+
<compilerPlugins>
217+
<compilerPlugin>
218+
<groupId>org.scalameta</groupId>
219+
<artifactId>semanticdb-scalac_2.12.10</artifactId>
220+
<version>4.3.24</version>
221+
</compilerPlugin>
222+
</compilerPlugins>
205223
</configuration>
206224
</plugin>
207225
<plugin>
@@ -225,6 +243,11 @@
225243
</scala>
226244
</configuration>
227245
</plugin>
246+
<plugin>
247+
<groupId>io.github.evis</groupId>
248+
<artifactId>scalafix-maven-plugin_2.13</artifactId>
249+
<version>0.1.8_0.11.0</version>
250+
</plugin>
228251
</plugins>
229252
</build>
230253
<packaging>jar</packaging>

spark/src/main/scala/com/alibaba/graphar/EdgeInfo.scala

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -16,8 +16,7 @@
1616

1717
package com.alibaba.graphar
1818

19-
import java.io.{File, FileInputStream}
20-
import org.apache.hadoop.fs.{Path, FileSystem}
19+
import org.apache.hadoop.fs.Path
2120
import org.apache.spark.sql.{SparkSession}
2221
import org.yaml.snakeyaml.{Yaml, DumperOptions}
2322
import org.yaml.snakeyaml.constructor.Constructor

spark/src/main/scala/com/alibaba/graphar/GraphInfo.scala

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -16,8 +16,7 @@
1616

1717
package com.alibaba.graphar
1818

19-
import java.io.{File, FileInputStream}
20-
import org.apache.hadoop.fs.{Path, FileSystem}
19+
import org.apache.hadoop.fs.Path
2120
import org.apache.spark.sql.{SparkSession}
2221
import org.yaml.snakeyaml.{Yaml, DumperOptions}
2322
import org.yaml.snakeyaml.constructor.Constructor

spark/src/main/scala/com/alibaba/graphar/VertexInfo.scala

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -16,8 +16,7 @@
1616

1717
package com.alibaba.graphar
1818

19-
import java.io.{File, FileInputStream}
20-
import org.apache.hadoop.fs.{Path, FileSystem}
19+
import org.apache.hadoop.fs.Path
2120
import org.apache.spark.sql.{SparkSession}
2221
import org.yaml.snakeyaml.{Yaml, DumperOptions}
2322
import org.yaml.snakeyaml.constructor.Constructor

spark/src/main/scala/com/alibaba/graphar/datasources/GarCommitProtocol.scala

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -21,9 +21,7 @@ import com.alibaba.graphar.GeneralParams
2121
import org.json4s._
2222
import org.json4s.jackson.JsonMethods._
2323

24-
import org.apache.spark.internal.io.FileCommitProtocol
2524
import org.apache.spark.sql.execution.datasources.SQLHadoopMapReduceCommitProtocol
26-
import org.apache.hadoop.mapreduce.task.TaskAttemptContextImpl
2725
import org.apache.hadoop.mapreduce._
2826
import org.apache.spark.internal.Logging
2927

spark/src/main/scala/com/alibaba/graphar/datasources/GarDataSource.scala

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,6 @@ import org.apache.spark.sql.SparkSession
2929
import org.apache.spark.sql.execution.datasources.csv.CSVFileFormat
3030
import org.apache.spark.sql.execution.datasources.orc.OrcFileFormat
3131
import org.apache.spark.sql.execution.datasources.parquet.ParquetFileFormat
32-
import org.apache.spark.sql.execution.datasources.v2._
3332
import org.apache.spark.sql.types.StructType
3433
import org.apache.spark.sql.util.CaseInsensitiveStringMap
3534
import org.apache.spark.sql.sources.DataSourceRegister

spark/src/main/scala/com/alibaba/graphar/datasources/GarWriterBuilder.scala

Lines changed: 2 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -26,12 +26,9 @@ import org.apache.hadoop.conf.Configuration
2626
import org.apache.hadoop.fs.Path
2727
import org.apache.hadoop.mapreduce.Job
2828
import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
29-
import org.apache.hadoop.mapreduce.{Job, TaskAttemptContext}
29+
import org.apache.hadoop.mapreduce.Job
3030

31-
import org.apache.spark.sql.execution.datasources.{
32-
OutputWriter,
33-
OutputWriterFactory
34-
}
31+
import org.apache.spark.sql.execution.datasources.OutputWriterFactory
3532
import org.apache.spark.sql.SparkSession
3633
import org.apache.spark.sql.catalyst.InternalRow
3734
import org.apache.spark.sql.catalyst.util.{CaseInsensitiveMap, DateTimeUtils}
@@ -49,7 +46,6 @@ import org.apache.spark.sql.execution.datasources.{
4946
import org.apache.spark.sql.execution.metric.SQLMetric
5047
import org.apache.spark.sql.internal.SQLConf
5148
import org.apache.spark.sql.types.{DataType, StructType}
52-
import org.apache.spark.sql.util.SchemaUtils
5349
import org.apache.spark.util.SerializableConfiguration
5450
import org.apache.spark.sql.execution.datasources.v2.FileBatchWrite
5551
import org.apache.spark.sql.catalyst.expressions.AttributeReference

spark/src/main/scala/com/alibaba/graphar/datasources/csv/CSVWriterBuilder.scala

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -30,8 +30,6 @@ import org.apache.spark.sql.execution.datasources.{
3030
import org.apache.spark.sql.execution.datasources.csv.CsvOutputWriter
3131
import org.apache.spark.sql.internal.SQLConf
3232
import org.apache.spark.sql.types.{DataType, StructType}
33-
import org.apache.spark.sql.sources.Filter
34-
import org.apache.spark.sql.connector.write.SupportsOverwrite
3533

3634
import com.alibaba.graphar.datasources.GarWriteBuilder
3735

spark/src/main/scala/com/alibaba/graphar/example/GraphAr2Neo4j.scala

Lines changed: 0 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -16,18 +16,11 @@
1616

1717
package com.alibaba.graphar.example
1818

19-
import com.alibaba.graphar.datasources._
20-
import com.alibaba.graphar.reader.{VertexReader, EdgeReader}
2119
import com.alibaba.graphar.graph.GraphReader
2220
import com.alibaba.graphar.{GeneralParams, GraphInfo}
2321
import com.alibaba.graphar.util.Utils
2422

25-
import java.io.{File, FileInputStream}
26-
import org.yaml.snakeyaml.Yaml
27-
import org.yaml.snakeyaml.constructor.Constructor
28-
import scala.beans.BeanProperty
2923
import org.apache.spark.sql.{DataFrame, SaveMode, SparkSession}
30-
import org.apache.hadoop.fs.{Path, FileSystem}
3124

3225
object GraphAr2Neo4j {
3326

0 commit comments

Comments
 (0)