clear query| facets| time Search criteria: author:"Josh Rosen".   Results from 41 to 50 from 586 (0.0s).
Loading phrases to help you
refine your search...
[SPARK-749] spark-ec2 fails to detect cluster after ssh error during launch - Spark - [issue]
...I tried to launch an EC2 cluster using Patrick's new version of the EC2 script, running this command ./spark-ec2 -i ~/.ssh/id_rsa -s 10 -t m1.large --spot-price 0.25 --region us-west-1 ...    Author: Josh Rosen , 2013-09-19, 14:58
[SPARK-750] LocalSparkContext should be included in Spark JAR - Spark - [issue]
...To aid third-party developers in writing unit tests with Spark, LocalSparkContext should be included in the Spark JAR.  Right now, it appears to be excluded because it is located in one...    Author: Josh Rosen , 2015-01-20, 15:46
[SPARK-885] PySpark shell should capture ctrl-c to prevent users from accidentally killing the Java gateway - Spark - [issue]
...If users hit 'ctrl-c' to discard a line of input (or interrupt a command) in the `pyspark` shell, the resulting KeyboardInterrupt kills the Py4J Java Gateway.  Maybe we could capture th...    Author: Josh Rosen , 2013-09-01, 15:10
[SPARK-919] spark-ec2 launch --resume doesn't re-initialize all modules - Spark - [issue]
...I launched a Spark cluster using the new EC2 scripts, stopped it with stop, then restarted it with start and ran launch --resume to re-deploy the Spark configurations.It looks like the scrip...    Author: Josh Rosen , 2013-10-04, 21:28
[SPARK-922] Update Spark AMI to Python 2.7 - Spark - [issue]
...Many Python libraries only support Python 2.7+, so we should make Python 2.7 the default Python on the Spark AMIs....    Author: Josh Rosen , 2016-04-22, 16:42
[SPARK-923] Bytes columns in web UI tables don't sort properly - Spark - [issue]
...In the Web UI, columns displaying sizes in bytes do not sort correctly; they are sorted by the string value instead of the number of bytes, megabytes, etc.  It's possible to add a sort ...    Author: Josh Rosen , 2013-11-30, 15:59
[SPARK-927] PySpark sample() doesn't work if numpy is installed on master but not on workers - Spark - [issue]
...PySpark's sample() method crashes with ImportErrors on the workers if numpy is installed on the driver machine but not on the workers.  I'm not sure what's the best way to fix this. &nb...    Author: Josh Rosen , 2015-01-05, 23:05
[SPARK-815] PySpark's parallelize() should batch objects after partitioning (instead of before) - Spark - [issue]
...PySpark uses batching when serializing and deserializing Python objects.  By default, it serializes objects in groups of 1024.The current batching code causes SparkContext.parallelize()...    Author: Josh Rosen , 2014-03-30, 23:33
[SPARK-822] defaultMinSplits can't be set higher than 2 - Spark - [issue]
...SparkContext.defaultMinSplits is used to control the default level of parallelism when reading input from files.  Unfortunately, the current implementation doesn't allow defaultMinSplit...    Author: Josh Rosen , 2015-01-25, 19:30
[SPARK-823] spark.default.parallelism's default is inconsistent across scheduler backends - Spark - [issue]
...The 0.7.3 configuration guide says that spark.default.parallelism's default is 8, but the default is actually max(totalCoreCount, 2) for the standalone scheduler backend, 8 for the Mesos sch...    Author: Josh Rosen , 2015-02-09, 16:31