Hi is this the right way to do that ??
```
spark
print(com.tr.dss.version)
import org.apache.spark.{SparkContext, SparkConf}
sc.stop()
val conf = new SparkConf()
.set("spark.executor.memory", "4g")
.set("spark.serializer", "org.apache.spark.serializer.KryoSerializer")
conf.registerKryoClasses(Array(classOf[dss.schema.logrecord]))
val sc = new SparkContext(conf)
print(sc.version)
```
N> dss.schema.logrecord is just avro generated class I am working on
I am getting
java.lang.IllegalStateException: SparkContext has been shutdown
further down In the code
even if
print(sc.version)
works ...