Products
Cloud
Sematext Cloud running on AWS infrastructure
Enterprise
Sematext Cloud running on your infrastructure
Infrastructure Monitoring
Infrastructure, application, container monitoring and alerting
Real User Monitoring
Get Invitation
Enhnance your site performance with data from actual site visitors
Logsene
Log Management â hosted ELK stack in the cloud
APM / Tracing
Get Invitation
Optimize end-to-end application performance
Product Updates
New
Check out whatâs new on Sematext Cloud
Sematext Docs
Sematext Documentation
Main Capabilities
Infrastructure Monitoring
Application Performance Monitoring
Log Management
Container Monitoring
Metrics and Logs
Alerts
Events
Integrations
API
Open Source Agents
Docker Agent
Kubernetes Agent
Logagent
Infrastructure Monitoring
Logsene Log Management
Docker Agent
Kubernetes Agent
Sematext Cloud
Logagent
Services
Consulting
Consulting Overview
Solr & Lucene Consulting
Elasticsearch Consulting
Logging Consulting
Support
Support Overview
Solr & Lucene Support
Elasticsearch Support
Training
Training Overview
Solr Training
Elasticsearch Training
Resources
Blog
About
Company
Customers
Case Studies
Partners
Testimonials
Jobs
Engineering
Contact
project
Spark
(586)
author
Josh Rosen
(586)
type
issue
(586)
mail # dev
(192)
mail # user
(140)
Apache
CNCF
DevOps
Containers
Logging
Monitoring
Automation
Machine Learning
Security
VIS
Database
Search
Big Data
1d
1w
1m
1y
all
time-biased relevance
relevancy
newest on top
oldest on top
Search
clear
query
|
facets
|
time
Search criteria:
author:"Josh Rosen"
. Results from
21
to
30
from
586
(0.0s).
Loading phrases to help you
refine your search...
[SPARK-1065] PySpark runs out of memory with large broadcast variables
-
Spark
- [issue]
...PySpark's driver components may run out of memory when broadcasting large variables (say 1 gigabyte).Because PySpark's broadcast is implemented on top of Java Spark's broadcast by broadcasti...
http://issues.apache.org/jira/browse/SPARK-1065
Author:
Josh Rosen
, 2014-08-17, 00:00
[SPARK-759] Change how we track AMI ids in the EC2 scripts
-
Spark
- [issue]
...I think we should change how we track AMI ids in the EC2 scripts.I don't like the current approach of using a URL to track the latest AMI id for each major version number:1. There's no versi...
http://issues.apache.org/jira/browse/SPARK-759
Author:
Josh Rosen
, 2013-08-26, 10:08
[SPARK-764] Fix SPARK_EXAMPLES_JAR in 0.7.2
-
Spark
- [issue]
...Users are reporting that SPARK_EXAMPLES_JAR is not set right in Spark 0.7.2 (see https://groups.google.com/d/msg/spark-users/nQ6wB2lcFN8/gWfBd6fLWHQJ for a recent example).A new post in http...
http://issues.apache.org/jira/browse/SPARK-764
Author:
Josh Rosen
, 2013-06-22, 10:22
[SPARK-776] Support adding jars to Spark shell
-
Spark
- [issue]
...We should add a mechanism to add additional jars to jobs run in the Spark shell, since addJar() doesn't work there (see https://github.com/mesos/spark/pull/359).There's a proposal/patch at h...
http://issues.apache.org/jira/browse/SPARK-776
Author:
Josh Rosen
, 2013-06-22, 17:17
[SPARK-778] run script should try java executable from JAVA_HOME first
-
Spark
- [issue]
...See https://groups.google.com/d/msg/spark-users/uuw2e9d1l74/_bCyDfq6dywJ;It looks like the script tries to run `java` first and falls back on `$JAVA_HOME/bin/java` only if `java` cannot be f...
http://issues.apache.org/jira/browse/SPARK-778
Author:
Josh Rosen
, 2013-06-27, 22:19
[SPARK-970] PySpark's saveAsTextFile() throws UnicodeEncodeError when saving unicode strings
-
Spark
- [issue]
...PySpark throws a UnicodeEncodeError when trying to save unicode objects to text files. This is because saveAsTextFile() calls str() to get objects' string representations, when it shou...
http://issues.apache.org/jira/browse/SPARK-970
Author:
Josh Rosen
, 2013-12-04, 11:10
[SPARK-971] Link to Confluence wiki from project website / documentation
-
Spark
- [issue]
...Spark's Confluence wiki (https://cwiki.apache.org/confluence/display/SPARK/Wiki+Homepage) is really hard to find; try a Google search for "apache spark wiki", for example.We should link to t...
http://issues.apache.org/jira/browse/SPARK-971
Author:
Josh Rosen
, 2014-11-10, 01:41
[SPARK-972] PySpark's "cannot run multiple SparkContexts at once" message should give source locations
-
Spark
- [issue]
...It can be difficult to debug PySpark's "Cannot run multiple SparkContexts at once" error message if you're not sure where the first context is being created; it would be helpful if the Spark...
http://issues.apache.org/jira/browse/SPARK-972
Author:
Josh Rosen
, 2016-07-12, 21:49
[SPARK-977] Add ZippedRDD / zip to PySpark
-
Spark
- [issue]
...We should add an equivalent of ZippedRDD / zip() to PySpark....
http://issues.apache.org/jira/browse/SPARK-977
Author:
Josh Rosen
, 2014-03-16, 22:30
[SPARK-985] Support Job Cancellation on Mesos Scheduler
-
Spark
- [issue]
...https://github.com/apache/incubator-spark/pull/29 added job cancellation but may still need support for Mesos scheduler backends:Quote: This looks good except that MesosSchedulerBackend isn'...
http://issues.apache.org/jira/browse/SPARK-985
Author:
Josh Rosen
, 2015-02-07, 22:46
< Prev
1
2
3
4
5
Next >