Thank you Andrew!

On Thu, Apr 5, 2018 at 7:17 AM, Andrew Otto <[email protected]> wrote:

> Tilman, aye ok thanks.  I will delay decom til next week.
>
> Chelsy yeah!  There should have been a hive-site.xml symlink created for
> spark2 when it was installed.  I’ll have to look into why that didn’t
> happen next time I poke around in the spark2 debian package, probably when
> we upgrade to spark 2.3.  In the meantime, I’ve manually created the
> symlink, so you should be able to SHOW DATABASES and see all your favorite
> friends now.
>
>
>
> On Tue, Apr 3, 2018 at 1:13 PM, Chelsy Xie <[email protected]> wrote:
>
>> Thanks Andrew and AE!!! ❤
>>
>> JupyterLab works like a charm! And I installed the R kernel under my
>> account, it works well too! 🎉
>>
>> Just one issue, I don't seem to be able to access Spark via notebook1003.
>> I can open a pyspark shell in the terminal on JupyterLab, but `SHOW
>> DATABASES` returns 'default', instead of a list of databases on hadoop.
>>
>> Chelsy
>>
>> On Mon, Apr 2, 2018 at 8:31 PM, Tilman Bayer <[email protected]>
>> wrote:
>>
>>> Thanks Andrew! I double-checked my folder and verified that all notebook
>>> files were copied over correctly.
>>>
>>> However, it is also worth mentioning that the kernel state is not
>>> preserved after the transfer, i.e. all running notebooks are stopped.
>>> This means for example that any variable values (say query results that are
>>> stored in a dataframe) will need to be recalculated or restored from e.g. a
>>> CSV or pickle file. It's good practice to save important data in that form
>>> anyway (notebooks can stop running for other reasons too, although they
>>> have usually stayed live for many days or weeks). Still, I can see an
>>> argument for holding off the decommissioning just a little longer, until
>>> say early next week, if that doesn't disrupt other things.
>>>
>>> In any case, +1 to what Leila said - I really appreciate the technical
>>> support for SWAP and am excited about the additional possibilities that
>>> this upgrade is bringing.
>>>
>>> On Mon, Apr 2, 2018 at 7:29 AM, Andrew Otto <[email protected]> wrote:
>>>
>>>> I have just done a final rsync of home directories from notebook1001
>>>> over to notebook1003.
>>>>
>>>> *Do not use notebook1001 anymore.*
>>>>
>>>> I will leave notebook1001 only for another day in case there or issues,
>>>> but plan to start the decom process this week.
>>>>
>>>> Thanks all!
>>>>
>>>>
>>>>
>>>> On Thu, Mar 22, 2018 at 3:34 PM, Andrew Otto <[email protected]>
>>>> wrote:
>>>>
>>>>> Hi everyone!
>>>>>
>>>>> *tl;dr stop using notebook1001 by Monday April 2nd, use notebook1003
>>>>> instead.*
>>>>>
>>>>> *(If you don’t have production access, you can ignore this email.)*
>>>>>
>>>>> As part of https://phabricator.wikimedia.org/T183145, we’ve ordered
>>>>> new hardware to replace the aging notebook1001.  The new servers are ready
>>>>> to go, so we need to schedule a deprecation timeline for notebook1001.
>>>>> That timeline is Monday April 2nd.  After that, your work on notebook1001
>>>>> will not longer be accessible.  Instead you should use notebook1003 (or
>>>>> notebook1004).
>>>>>
>>>>> But there is good news too!  Last week I rsynced everyone’s home
>>>>> directories from notebook1001 over to notebook1003.  I also upgraded the
>>>>> default virtualenv your notebooks run from.  Your notebook files should 
>>>>> all
>>>>> be accessible on notebook1003.  However, the version of Python3 changed
>>>>> from 3.4 to 3.5 during this upgrade.  Dependencies that your notebook uses
>>>>> that you installed on notebook1001 may not be available at first.  You
>>>>> might need to redo a pip install those dependencies into the new notebook
>>>>> Python 3.5 virtualenv.  (I can’t really give you explicit instructions to
>>>>> do that, as I don’t know what you use for your notebooks.)
>>>>>
>>>>> I’ll do a final rsync any newer files in home directories from
>>>>> notebook1001 on Monday April 2nd.  If you’ve been working on notebook1001
>>>>> since after March 15th, this should get everything up to date on
>>>>> notebook1003 before notebook1001 goes away.  BUT!  *Do not work on
>>>>> both notebook1001 and notebook1003*!  My final rsync will keep the
>>>>> most recently modified version of files from either server.
>>>>>
>>>>>
>>>>> OOooOo and there’s even more good news!  I’ve made the notebooks able
>>>>> to access system site packages, and installed a ton of useful packages
>>>>> <https://github.com/wikimedia/puppet/blob/production/modules/statistics/manifests/packages.pp#L77-L98>
>>>>> by default.  pandas, scipy, requests, etc.  If there’s something else you
>>>>> think you might need, let us know.  Or just pip install it into your
>>>>> notebook.
>>>>>
>>>>> Additionally, pyhive has been installed too, so you should be able to
>>>>> more easily access Hive directly from a python notebook.
>>>>>
>>>>> I’ve updated docs at https://wikitech.wikimedia.org/wiki/SWAP#Usage,
>>>>> please take a look.
>>>>>
>>>>> If you have any questions, please don’t hesitate to ask, either here
>>>>> on or phabricator: https://phabricator.wikimedia.org/T183145.
>>>>>
>>>>> - Andrew Otto & Analytics Engineering
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>
>>>
>>>
>>> --
>>> Tilman Bayer
>>> Senior Analyst
>>> Wikimedia Foundation
>>> IRC (Freenode): HaeB
>>>
>>
>>
>>
>> --
>> *Chelsy Xie*
>> Data Analyst
>> Wikimedia Foundation
>>
>
>


-- 
*Chelsy Xie*
Data Analyst
Wikimedia Foundation
_______________________________________________
Analytics mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/analytics

Reply via email to