Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Branch 2.2 merge #230

Merged
merged 7 commits into from
May 1, 2018
Merged

Commits on Apr 9, 2018

  1. [SPARK-23816][CORE] Killed tasks should ignore FetchFailures.

    SPARK-19276 ensured that FetchFailures do not get swallowed by other
    layers of exception handling, but it also meant that a killed task could
    look like a fetch failure.  This is particularly a problem with
    speculative execution, where we expect to kill tasks as they are reading
    shuffle data.  The fix is to ensure that we always check for killed
    tasks first.
    
    Added a new unit test which fails before the fix, ran it 1k times to
    check for flakiness.  Full suite of tests on jenkins.
    
    Author: Imran Rashid <irashid@cloudera.com>
    
    Closes apache#20987 from squito/SPARK-23816.
    
    (cherry picked from commit 10f45bb)
    Signed-off-by: Marcelo Vanzin <vanzin@cloudera.com>
    squito authored and Marcelo Vanzin committed Apr 9, 2018
    Configuration menu
    Copy the full SHA
    e957c4e View commit details
    Browse the repository at this point in the history

Commits on Apr 10, 2018

  1. Configuration menu
    Copy the full SHA
    78215fa View commit details
    Browse the repository at this point in the history

Commits on Apr 18, 2018

  1. [SPARK-24007][SQL] EqualNullSafe for FloatType and DoubleType might g…

    …enerate a wrong result by codegen.
    
    `EqualNullSafe` for `FloatType` and `DoubleType` might generate a wrong result by codegen.
    
    ```scala
    scala> val df = Seq((Some(-1.0d), None), (None, Some(-1.0d))).toDF()
    df: org.apache.spark.sql.DataFrame = [_1: double, _2: double]
    
    scala> df.show()
    +----+----+
    |  _1|  _2|
    +----+----+
    |-1.0|null|
    |null|-1.0|
    +----+----+
    
    scala> df.filter("_1 <=> _2").show()
    +----+----+
    |  _1|  _2|
    +----+----+
    |-1.0|null|
    |null|-1.0|
    +----+----+
    ```
    
    The result should be empty but the result remains two rows.
    
    Added a test.
    
    Author: Takuya UESHIN <ueshin@databricks.com>
    
    Closes apache#21094 from ueshin/issues/SPARK-24007/equalnullsafe.
    
    (cherry picked from commit f09a9e9)
    Signed-off-by: gatorsmile <gatorsmile@gmail.com>
    ueshin authored and gatorsmile committed Apr 18, 2018
    Configuration menu
    Copy the full SHA
    a902323 View commit details
    Browse the repository at this point in the history
  2. [SPARK-23963][SQL] Properly handle large number of columns in query o…

    …n text-based Hive table
    
    ## What changes were proposed in this pull request?
    
    TableReader would get disproportionately slower as the number of columns in the query increased.
    
    I fixed the way TableReader was looking up metadata for each column in the row. Previously, it had been looking up this data in linked lists, accessing each linked list by an index (column number). Now it looks up this data in arrays, where indexing by column number works better.
    
    ## How was this patch tested?
    
    Manual testing
    All sbt unit tests
    python sql tests
    
    Author: Bruce Robbins <bersprockets@gmail.com>
    
    Closes apache#21043 from bersprockets/tabreadfix.
    bersprockets authored and gatorsmile committed Apr 18, 2018
    Configuration menu
    Copy the full SHA
    041aec4 View commit details
    Browse the repository at this point in the history

Commits on Apr 24, 2018

  1. [MINOR][DOCS] Fix comments of SQLExecution#withExecutionId

    ## What changes were proposed in this pull request?
    Fix comment. Change `BroadcastHashJoin.broadcastFuture` to `BroadcastExchangeExec.relationFuture`: https://github.com/apache/spark/blob/d28d5732ae205771f1f443b15b10e64dcffb5ff0/sql/core/src/main/scala/org/apache/spark/sql/execution/exchange/BroadcastExchangeExec.scala#L66
    
    ## How was this patch tested?
    N/A
    
    Author: seancxmao <seancxmao@gmail.com>
    
    Closes apache#21113 from seancxmao/SPARK-13136.
    
    (cherry picked from commit c303b1b)
    Signed-off-by: hyukjinkwon <gurwls223@apache.org>
    seancxmao authored and HyukjinKwon committed Apr 24, 2018
    Configuration menu
    Copy the full SHA
    e77d62a View commit details
    Browse the repository at this point in the history

Commits on Apr 27, 2018

  1. Configuration menu
    Copy the full SHA
    cbb272e View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    6875599 View commit details
    Browse the repository at this point in the history