clear query| facets| time Search criteria: author:"Yin Huai".   Results from 11 to 20 from 390 (0.0s).
Loading phrases to help you
refine your search...
[SPARK-10737] When using UnsafeRows, SortMergeJoin may return wrong results - Spark - [issue]
...val df1 = (1 to 10).map(i => (s"str_$i", i)).toDF("i", "j")val df2 =  df1  .join(df1.select(df1("i")), "i")  .select(df1("i"), df1("j"))val df3 = df2.withColumnRenamed("i",...
http://issues.apache.org/jira/browse/SPARK-10737    Author: Yin Huai , 2015-09-22, 20:32
[SPARK-6016] Cannot read the parquet table after overwriting the existing table when spark.sql.parquet.cacheMetadata=true - Spark - [issue]
...saveAsTable is fine and seems we have successfully deleted the old data and written the new data. However, when reading the newly created table, an error will be thrown.Error in SQL statemen...
http://issues.apache.org/jira/browse/SPARK-6016    Author: Yin Huai , 2015-04-24, 00:44
[SPARK-6023] ParquetConversions fails to replace the destination MetastoreRelation of an InsertIntoTable node to ParquetRelation2 - Spark - [issue]
...import sqlContext._sql("drop table if exists test")val df1 = sqlContext.jsonRDD(sc.parallelize((1 to 10).map(i => s"""{"a":$i}""")))df1.registerTempTable("jt")sql("create table test (a bi...
http://issues.apache.org/jira/browse/SPARK-6023    Author: Yin Huai , 2015-04-24, 00:43
[SPARK-6024] When a data source table has too many columns, it's schema cannot be stored in metastore. - Spark - [issue]
...Because we are using table properties of a Hive metastore table to store the schema, when a schema is too wide, we cannot persist it in metastore.15/02/25 18:13:50 ERROR metastore.RetryingHM...
http://issues.apache.org/jira/browse/SPARK-6024    Author: Yin Huai , 2015-02-27, 04:46
[SPARK-6052] In JSON schema inference, we should always set containsNull of an ArrayType to true - Spark - [issue]
...We should not try to figure out if an array contains null or not because we may miss arrays with null if we do sampling or future data may have nulls in the array....
http://issues.apache.org/jira/browse/SPARK-6052    Author: Yin Huai , 2015-04-24, 00:42
[SPARK-6073] Need to refresh metastore cache after append data in CreateMetastoreDataSourceAsSelect - Spark - [issue]
...We should drop the metadata cache in CreateMetastoreDataSourceAsSelect after we append data. Otherwise, users have to manually call HiveContext.refreshTable to drop the cached metadata entry...
http://issues.apache.org/jira/browse/SPARK-6073    Author: Yin Huai , 2015-04-24, 00:42
[SPARK-6123] Parquet reader should use the schema of every file to create converter - Spark - [issue]
...For two parquet files for the same table having an array column, if values of the array in one file was created when containsNull was true and those in another file was created when contains...
http://issues.apache.org/jira/browse/SPARK-6123    Author: Yin Huai , 2015-07-09, 11:36
[SPARK-6146] Support more datatype in SqlParser - Spark - [issue]
...Right now, I cannot do df.selectExpr("cast(a as bigint)")because only the following data types are supported in SqlParserprotected lazy val dataType: Parser[DataType] =    ( STRING...
http://issues.apache.org/jira/browse/SPARK-6146    Author: Yin Huai , 2015-04-24, 00:39
[SPARK-5758] Use LongType as the default type for integers in JSON schema inference. - Spark - [issue]
...Per discussion in https://github.com/apache/spark/pull/4521, we will use LongType as the default data type for integer values in JSON schema inference....
http://issues.apache.org/jira/browse/SPARK-5758    Author: Yin Huai , 2015-04-25, 21:46
[SPARK-5781] Add metadata files for JSON datasets - Spark - [issue]
...If we save a dataset in JSON format (e.g. through DataFrame.save), we should also persist the schema of the table. So, we can avoid inferring the schema when we want to query it in future....
http://issues.apache.org/jira/browse/SPARK-5781    Author: Yin Huai , 2016-10-07, 22:43