I'm currently running Zeppelin on a spark master node using the AWS
provided Zeppelin install. I'm trying to get the notebook setup so
multiple devs can use it (and the spark cluster) concurrently. I have
the spark interpreter set to instantiate 'Per Note' in 'isolated'
processes. I also have 'spark.dynamicAllocation.enabled' set to 'true'
so the multiple spark contexts can share the cluster.
The problem I'm seeing is when the second spark context tries to
instantiate hive starts throwing errors because the Derby database has
already been booted (by the other context). Full stack trace is
available here . How do I go about working around this? Is there a
way to have it use another database or is this a limitation?
Thanks for any help!
Austin L. Heyne