Github user HyukjinKwon commented on the issue:

    https://github.com/apache/spark/pull/19674
  
    @ueshin, now I am getting what you meant by DST. I think you roughly knew 
this problem but let me describe it more given my debugging.
    
    Looks the problem is in `mktime` and sounds `mktime` is platform-dependent 
(roughly assuming from codes and docs).
    
    I made a minimised reproducer:
    
    ```python
    import time
    import os
    from datetime import datetime
    os.environ["TZ"] = "America/Los_Angeles"
    time.tzset()
    time.mktime(datetime(2100, 4, 4, 4, 4, 4).timetuple())
    ```
    
    My local, it prints: 
    
    ```
    4110523444.0
    ```
    
    On Unbuntu 14.04:
    
    ```
    4110519844.0
    ```
    
    Jenkins, it prints:
    
    ```
    4110519844.0
    ```
    
    I am not sure if this is easily fixable within Spark as it looks dependent 
on Python implementation and/or the underlying C library, up to my knowledge 
and from my reading some docs.
    
    Could you maybe avoid this time within DST for now in your PR? I currently 
don't have a good idea to fix this with a simple and surgical fix


---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to