Subject: org.apache.spark.util.SparkUncaughtExceptionHandler

Hi Nimmi,

Can you send us the spark parameters with overhead. assuming you are
running with yarn


[4] - 864GB

--num-executors 32

--executor-memory 21G

--executor-cores 4
--conf spark.yarn.executor.memoryOverhead=3000

 The parameter spark.yarn.executor.memoryOverhead is explained as below:

 spark.yarn.executor.memoryOverhead = executorMemory * 0.10, with minimum
of 384

 The amount of off-heap memory (in megabytes) to be allocated per executor.
This is memory that accounts for things like VM overheads, interned
strings, other native overheads, etc. This tends to grow with the executor
size (typically


Dr Mich Talebzadeh

LinkedIn *
*Disclaimer:* Use it at your own risk. Any and all responsibility for any
loss, damage or destruction of data or any other property which may arise
from relying on this email's technical content is explicitly disclaimed.
The author will in no case be liable for any monetary damages arising from
such loss, damage or destruction.
On Thu, 10 Oct 2019 at 21:39, Nimmi Cv <[EMAIL PROTECTED]> wrote: