clear query| facets| time Search criteria: .   Results from 1 to 10 from 49 (0.0s).
Loading phrases to help you
refine your search...
[SPARK-30282] Migrate SHOW TBLPROPERTIES to new framework - Spark - [issue]
...For the following v2 commands, Analyzer.ResolveTables does not check against the temp views before resolving UnresolvedV2Relation, thus it always resolves UnresolvedV2Relation to a table: AL...    Author: Terry Kim , 2020-07-13, 18:26
[SPARK-32282] Improve EnsureRquirement.reorderJoinKeys to handle more scenarios such as PartitioningCollection - Spark - [issue]
...The EnsureRquirement.reorderJoinKeys can be improved to handle the following scenarios: If the keys cannot be reordered to match the left-side HashPartitioning, consider the right-side HashP...    Author: Terry Kim , 2020-07-12, 03:11
[SPARK-31625] Unregister application from YARN resource manager outside the shutdown hook - Spark - [issue]
...Currently, an application is unregistered from YARN resource manager as a shutdown hook. In the scenario where the shutdown hook does not run (e.g., timeouts, etc.), the application is not u...    Author: Terry Kim , 2020-07-07, 20:53
[VOTE] Decommissioning SPIP - Spark - [mail # dev]
...+1 (non-binding)Thanks,TerryOn Wed, Jul 1, 2020 at 6:05 PM Holden Karau  wrote:> Hi Spark Devs,>> I think discussion has settled on the SPIP doc at>
   Author: Terry Kim , 2020-07-06, 17:17
Announcing .NET for Apache Spark™ 0.12 - Spark - [mail # user]
...We are happy to announce that .NET for Apache Spark™ v0.12 has been released! Thanks to the community for thegreat feedback. The release noteincludes the full list of features/improvements o...
   Author: Terry Kim , 2020-07-02, 19:18
Hyperspace v0.1 is now open-sourced! - Spark - [mail # user]
...Hi all,We are happy to announce the open-sourcing of Hyperspace v0.1, an indexingsubsystem for Apache Spark™:   - Code:   - Blog A...
   Author: Terry Kim , 2020-07-02, 17:56
[SPARK-31350] Coalesce bucketed tables for join if applicable - Spark - [issue]
...The following example of joining two bucketed tables introduces a full shuffle:spark.conf.set("spark.sql.autoBroadcastJoinThreshold", "0")val df1 = (0 until 20).map(i => (i % 5, i % 13, i...    Author: Terry Kim , 2020-06-19, 23:22
[SPARK-30065] Unable to drop na with duplicate columns - Spark - [issue]
...Trying to drop rows with null values fails even when no columns are specified. This should be allowed:scala> val left = Seq(("1", null), ("3", "4")).toDF("col1", "col2")left: org.apache.s...    Author: Terry Kim , 2020-06-08, 16:14
[expand - 1 more] - Using existing distribution for join when subset of keys - Spark - [mail # user]
...Is the following what you trying to do?spark.conf.set("spark.sql.autoBroadcastJoinThreshold", "0")val df1 = (0 until 100).map(i => (i % 5, i % 13)).toDF("x", "y")val df2 = (0 until 100).m...
   Author: Terry Kim , 2020-05-31, 23:14
[SPARK-31869] BroadcastHashJoinExe's outputPartitioning can utilize the build side - Spark - [issue]
...Currently, the BroadcastHashJoinExec's outputPartitioning only uses the streamed side's outputPartitioning. Thus, if the join key is from the build side for the join where one side is Broadc...    Author: Terry Kim , 2020-05-30, 04:21