It is a bug. You should not have to do this. Thanks for trying it out
and letting me know.

---------
Matt Taylor
OS Community Flag-Bearer
Numenta


On Mon, Feb 23, 2015 at 11:21 AM, David Wood <[email protected]> wrote:
> Hi Matt,
>
> Yes, thanks.
>
> I can run several of the remaining examples (at least) by setting the NUPIC 
> and NTA_DATA_PATH environment variables. It would be handy if that was 
> documented in a README.
>
> Regards,
> Dave
> --
> http://about.me/david_wood
>
>
>
>> On Feb 22, 2015, at 12:40, Matthew Taylor <[email protected]> wrote:
>>
>> Thanks again. I have a bug report for that problem now:
>> https://github.com/numenta/nupic/issues/1805
>>
>> Does that look familiar?
>> ---------
>> Matt Taylor
>> OS Community Flag-Bearer
>> Numenta
>>
>>
>> On Sun, Feb 22, 2015 at 9:22 AM, David Wood <[email protected]> wrote:
>>> Thanks for your quick response, Matt.
>>>
>>>> On Feb 21, 2015, at 23:46, Matthew Taylor <[email protected]> wrote:
>>>>
>>>> Hi David, thanks for the report. Unfortunately, I cannot tell what the
>>>> root error is, because of
>>>> https://github.com/numenta/nupic/issues/1815.
>>>
>>>
>>> Oh, yes, that looks familiar.  Sorry I didn’t see that before reporting.
>>>
>>>
>>>> If you don't mind doing a little legwork, you might be able to help us
>>>> identify the problem.
>>>>
>>>> 1. Download our source code by either:
>>>> - git clone https://github.com/numenta/nupic.git
>>>> - or get the tarball at https://github.com/numenta/nupic/archive/master.zip
>>>> 2. Run the simple hotgym example code:
>>>> - cd nupic/examples/opf/clients/hotgym/simple
>>>> - python hotgym.py
>>>>
>>>> Does this give you any errors?
>>>
>>>
>>> Yes, the error was:
>>> [[
>>> Traceback (most recent call last):
>>>  File "hotgym.py", line 94, in <module>
>>>    runHotgym()
>>>  File "hotgym.py", line 69, in runHotgym
>>>    with open (findDataset(_DATA_PATH)) as fin:
>>>  File 
>>> "/usr/local/lib/python2.7/dist-packages/nupic/data/datasethelpers.py", line 
>>> 79, in findDataset
>>>    (datasetPath, os.environ.get('NTA_DATA_PATH', '')))
>>> Exception: Unable to locate: extra/hotgym/rec-center-hourly.csv using 
>>> NTA_DATA_PATH of
>>> ]]
>>>
>>> The last line was really truncated as shown. There was nothing following 
>>> the trailing “of” other than the shell prompt.
>>>
>>> So the problem does seem to be related to the data file not being found. I 
>>> changed this line in hotgym.py:
>>> [[
>>> _DATA_PATH = "extra/hotgym/rec-center-hourly.csv”
>>> ]]
>>> …to use an absolute file path, which avoids the problem. The script then 
>>> runs fine.
>>>
>>> I haven’t dug deeply enough yet to see where the path problem is (my Python 
>>> skills need some dusting), but it is clearly not resolving the relative 
>>> path correctly.
>>>
>>> Regards,
>>> Dave
>>> --
>>> http://about.me/david_wood
>>>
>>>
>>>>
>>>> Thanks,
>>>> ---------
>>>> Matt Taylor
>>>> OS Community Flag-Bearer
>>>> Numenta
>>>>
>>>>
>>>> On Sat, Feb 21, 2015 at 5:43 PM, David Wood <[email protected]> wrote:
>>>>> Hi all,
>>>>>
>>>>> I have recently read Jeff’s book and the Numenta white paper, and am very
>>>>> interested in getting started with NuPIC. Unfortunately, I am having some
>>>>> trouble getting a clean installation.
>>>>>
>>>>> My environment is a Rackspace cloud VM running:
>>>>> - Ubuntu 14.10 (Utopic Unicorn)
>>>>> - Python 2.7.8
>>>>> - mysql  Ver 14.14 Distrib 5.6.19, for debian-linux-gnu (x86_64) using
>>>>> EditLine wrapper
>>>>>
>>>>> I installed NuPIC as follows:
>>>>> [[
>>>>> # apt-get install python-pip
>>>>> # apt-get install python-dev
>>>>> # pip install numpy
>>>>> # pip install
>>>>> https://s3-us-west-2.amazonaws.com/artifacts.numenta.org/numenta/nupic/releases/nupic-0.1.3-cp27-none-linux_x86_64.whl
>>>>> ]]
>>>>>
>>>>> …and then attempted to run the “hot gym” predication example:
>>>>> [[
>>>>> # cd ~/nupic/examples/opf/clients/hotgym/prediction/one_gym
>>>>> # ./swarm.py
>>>>> ]]
>>>>>
>>>>> The output of the swarm run is provided below my signature.
>>>>>
>>>>> Unfortunately, the numpy and nupic installations as well as the hotgym
>>>>> example as seem to have significant numbers of python errors. The errors 
>>>>> are
>>>>> both trivial and apparently problematic, such as API changes. The swarm 
>>>>> run
>>>>> shows, for example, a TypeError.
>>>>>
>>>>> Should I expect this example to “just work” in the 0.13 release, or is my
>>>>> environment too new? Would a downgrade of Ubuntu or Python “fix” the
>>>>> problem? Does anyone have other suggestions? Thanks in advance!
>>>>>
>>>>> Regards,
>>>>> Dave
>>>>> --
>>>>> http://about.me/david_wood
>>>>>
>>>>>
>>>>> The output of the swarm run was:
>>>>> [[
>>>>> This script runs a swarm on the input data (rec-center-hourly.csv) and
>>>>> creates a model parameters file in the `model_params` directory containing
>>>>> the best model found by the swarm. Dumps a bunch of crud to stdout because
>>>>> that is just what swarming does at this point. You really don't need to
>>>>> pay any attention to it.
>>>>>
>>>>> =================================================
>>>>> = Swarming on rec-center-hourly data...
>>>>> = Medium swarm. Sit back and relax, this could take awhile.
>>>>> =================================================
>>>>> Generating experiment files in directory:
>>>>> /root/nupic/examples/opf/clients/hotgym/prediction/one_gym/swarm...
>>>>> Writing  313 lines...
>>>>> Writing  113 lines...
>>>>> done.
>>>>> None
>>>>> Successfully submitted new HyperSearch job, jobID=1002
>>>>> Evaluated 0 models
>>>>> HyperSearch finished!
>>>>> Worker completion message: None
>>>>>
>>>>> Results from all experiments:
>>>>> ----------------------------------------------------------------
>>>>> Generating experiment files in directory: /tmp/tmpBvtweU...
>>>>> Writing  313 lines...
>>>>> Writing  113 lines...
>>>>> done.
>>>>> None
>>>>> json.loads(jobInfo.results) raised an exception.  Here is some info to 
>>>>> help
>>>>> with debugging:
>>>>> jobInfo:  _jobInfoNamedTuple(jobId=1002, client=u'GRP', clientInfo=u'',
>>>>> clientKey=u'', cmdLine=u'$HYPERSEARCH', params=u'{"hsVersion": "v2",
>>>>> "maxModels": null, "persistentJobGUID":
>>>>> "8090e46e-ba32-11e4-ad72-bc764e202244", "useTerminators": false,
>>>>> "description": {"includedFields": [{"fieldName": "timestamp", "fieldType":
>>>>> "datetime"}, {"maxValue": 53.0, "fieldName": "kw_energy_consumption",
>>>>> "fieldType": "float", "minValue": 0.0}], "streamDef": {"info":
>>>>> "kw_energy_consumption", "version": 1, "streams": [{"info": "Rec Center",
>>>>> "source": "file://rec-center-hourly.csv", "columns": ["*"]}]},
>>>>> "inferenceType": "TemporalMultiStep", "inferenceArgs": {"predictionSteps":
>>>>> [1], "predictedField": "kw_energy_consumption"}, "iterationCount": -1,
>>>>> "swarmSize": "medium"}}', jobHash='\x80\x90\xe4o\xba2\x11\xe4\xadr\xbcvN
>>>>> "D', status=u'notStarted', completionReason=None, completionMsg=None,
>>>>> workerCompletionReason=u'success', workerCompletionMsg=None, cancel=0,
>>>>> startTime=None, endTime=None, results=None, engJobType=u'hypersearch',
>>>>> minimumWorkers=1, maximumWorkers=4, priority=0, engAllocateNewWorkers=1,
>>>>> engUntendedDeadWorkers=0, numFailedWorkers=0, 
>>>>> lastFailedWorkerErrorMsg=None,
>>>>> engCleaningStatus=u'notdone', genBaseDescription=None, 
>>>>> genPermutations=None,
>>>>> engLastUpdateTime=datetime.datetime(2015, 2, 22, 1, 31, 19),
>>>>> engCjmConnId=None, engWorkerState=None, engStatus=None,
>>>>> engModelMilestones=None)
>>>>> jobInfo.results:  None
>>>>> EXCEPTION:  expected string or buffer
>>>>> Traceback (most recent call last):
>>>>> File "./swarm.py", line 109, in <module>
>>>>>   swarm(INPUT_FILE)
>>>>> File "./swarm.py", line 101, in swarm
>>>>>   modelParams = swarmForBestModelParams(SWARM_DESCRIPTION, name)
>>>>> File "./swarm.py", line 78, in swarmForBestModelParams
>>>>>   verbosity=0
>>>>> File
>>>>> "/usr/local/lib/python2.7/dist-packages/nupic/swarming/permutations_runner.py",
>>>>> line 276, in runWithConfig
>>>>>   return _runAction(runOptions)
>>>>> File
>>>>> "/usr/local/lib/python2.7/dist-packages/nupic/swarming/permutations_runner.py",
>>>>> line 217, in _runAction
>>>>>   returnValue = _runHyperSearch(runOptions)
>>>>> File
>>>>> "/usr/local/lib/python2.7/dist-packages/nupic/swarming/permutations_runner.py",
>>>>> line 160, in _runHyperSearch
>>>>>   metricsKeys=search.getDiscoveredMetricsKeys())
>>>>> File
>>>>> "/usr/local/lib/python2.7/dist-packages/nupic/swarming/permutations_runner.py",
>>>>> line 825, in generateReport
>>>>>   results = json.loads(jobInfo.results)
>>>>> File
>>>>> "/usr/local/lib/python2.7/dist-packages/nupic/support/object_json.py", 
>>>>> line
>>>>> 163, in loads
>>>>>   json.loads(s, object_hook=objectDecoderHook, **kwargs))
>>>>> File "/usr/lib/python2.7/json/__init__.py", line 351, in loads
>>>>>   return cls(encoding=encoding, **kw).decode(s)
>>>>> File "/usr/lib/python2.7/json/decoder.py", line 366, in decode
>>>>>   obj, end = self.raw_decode(s, idx=_w(s, 0).end())
>>>>> TypeError: expected string or buffer
>>>>> ]]
>>>>>
>>>>
>>>
>>>
>>
>
>

Reply via email to