Started by GitHub push by severinsimmler
Building remotely on Rechenknecht in workspace /mnt/data/jenkins/workspace/DARIAH-Topics
> git rev-parse --is-inside-work-tree # timeout=10
Fetching changes from the remote Git repository
> git config remote.origin.url https://github.com/DARIAH-DE/Topics # timeout=10
Fetching upstream changes from https://github.com/DARIAH-DE/Topics
> git --version # timeout=10
using GIT_ASKPASS to set credentials
> git fetch --tags --progress https://github.com/DARIAH-DE/Topics +refs/heads/*:refs/remotes/origin/*
> git rev-parse refs/remotes/origin/testing^{commit} # timeout=10
> git rev-parse refs/remotes/origin/origin/testing^{commit} # timeout=10
Checking out Revision b193131a17023fa0cc9b8d0795f2399dd77d48fe (refs/remotes/origin/testing)
> git config core.sparsecheckout # timeout=10
> git checkout -f b193131a17023fa0cc9b8d0795f2399dd77d48fe
> git rev-list d59973904afc052fbd745a85cf2b5cc24039dd22 # timeout=10
[DARIAH-Topics] $ /usr/bin/python3 /mnt/data/jenkins/shiningpanda/jobs/62c67c92/virtualenv.py /mnt/data/jenkins/shiningpanda/jobs/62c67c92/virtualenvs/d41d8cd9
Using base prefix '/usr'
New python executable in /mnt/data/jenkins/shiningpanda/jobs/62c67c92/virtualenvs/d41d8cd9/bin/python3
Also creating executable in /mnt/data/jenkins/shiningpanda/jobs/62c67c92/virtualenvs/d41d8cd9/bin/python
Installing setuptools, pip, wheel...done.
[DARIAH-Topics] $ /bin/sh -xe /tmp/shiningpanda1460602624174767729.sh
+ pip install -U pip
Requirement already up-to-date: pip in /mnt/data/jenkins/shiningpanda/jobs/62c67c92/virtualenvs/d41d8cd9/lib/python3.5/site-packages
+ pip install -U -r requirements-dev.txt
Obtaining file:///mnt/data/jenkins/workspace/DARIAH-Topics (from -r requirements.txt (line 1))
Collecting nose (from -r requirements-dev.txt (line 2))
Using cached nose-1.3.7-py3-none-any.whl
Collecting nosexcover (from -r requirements-dev.txt (line 3))
Using cached nosexcover-1.0.11-py2.py3-none-any.whl
Collecting jupyter (from -r requirements-dev.txt (line 4))
Using cached jupyter-1.0.0-py2.py3-none-any.whl
Collecting pandas>=0.19.2 (from dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Using cached pandas-0.19.2-cp35-cp35m-manylinux1_x86_64.whl
Collecting regex>=2017.01.14 (from dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Collecting gensim>=0.13.2 (from dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Collecting matplotlib==1.5.3 (from dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Using cached matplotlib-1.5.3-cp35-cp35m-manylinux1_x86_64.whl
Collecting numpy>=1.3 (from dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Using cached numpy-1.12.1-cp35-cp35m-manylinux1_x86_64.whl
Collecting scipy>=0.7 (from dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Using cached scipy-0.19.0-cp35-cp35m-manylinux1_x86_64.whl
Collecting werkzeug>=0.11.15 (from dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Using cached Werkzeug-0.12.1-py2.py3-none-any.whl
Collecting flask>=0.11.1 (from dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Using cached Flask-0.12.1-py2.py3-none-any.whl
Collecting wikipedia>=1.4.0 (from dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Collecting lxml>=3.6.4 (from dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Using cached lxml-3.7.3-cp35-cp35m-manylinux1_x86_64.whl
Collecting pyLDAvis>=2.0.0 (from dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Collecting coverage>=3.4 (from nosexcover->-r requirements-dev.txt (line 3))
Using cached coverage-4.3.4-cp35-cp35m-manylinux1_x86_64.whl
Collecting nbconvert (from jupyter->-r requirements-dev.txt (line 4))
Using cached nbconvert-5.1.1-py2.py3-none-any.whl
Collecting ipywidgets (from jupyter->-r requirements-dev.txt (line 4))
Using cached ipywidgets-6.0.0-py2.py3-none-any.whl
Collecting ipykernel (from jupyter->-r requirements-dev.txt (line 4))
Using cached ipykernel-4.6.1-py3-none-any.whl
Collecting qtconsole (from jupyter->-r requirements-dev.txt (line 4))
Using cached qtconsole-4.3.0-py2.py3-none-any.whl
Collecting jupyter-console (from jupyter->-r requirements-dev.txt (line 4))
Using cached jupyter_console-5.1.0-py2.py3-none-any.whl
Collecting notebook (from jupyter->-r requirements-dev.txt (line 4))
Using cached notebook-5.0.0-py2.py3-none-any.whl
Collecting pytz>=2011k (from pandas>=0.19.2->dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Using cached pytz-2017.2-py2.py3-none-any.whl
Collecting python-dateutil>=2 (from pandas>=0.19.2->dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Using cached python_dateutil-2.6.0-py2.py3-none-any.whl
Collecting smart-open>=1.2.1 (from gensim>=0.13.2->dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Requirement already up-to-date: six>=1.5.0 in /mnt/data/jenkins/shiningpanda/jobs/62c67c92/virtualenvs/d41d8cd9/lib/python3.5/site-packages (from gensim>=0.13.2->dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Requirement already up-to-date: pyparsing!=2.0.0,!=2.0.4,!=2.1.2,>=1.5.6 in /mnt/data/jenkins/shiningpanda/jobs/62c67c92/virtualenvs/d41d8cd9/lib/python3.5/site-packages (from matplotlib==1.5.3->dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Collecting cycler (from matplotlib==1.5.3->dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Using cached cycler-0.10.0-py2.py3-none-any.whl
Collecting itsdangerous>=0.21 (from flask>=0.11.1->dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Collecting Jinja2>=2.4 (from flask>=0.11.1->dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Using cached Jinja2-2.9.6-py2.py3-none-any.whl
Collecting click>=2.0 (from flask>=0.11.1->dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Using cached click-6.7-py2.py3-none-any.whl
Collecting beautifulsoup4 (from wikipedia>=1.4.0->dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Using cached beautifulsoup4-4.5.3-py3-none-any.whl
Collecting requests<3.0.0,>=2.0.0 (from wikipedia>=1.4.0->dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Using cached requests-2.13.0-py2.py3-none-any.whl
Collecting future (from pyLDAvis>=2.0.0->dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Collecting joblib>=0.8.4 (from pyLDAvis>=2.0.0->dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Using cached joblib-0.11-py2.py3-none-any.whl
Collecting pytest (from pyLDAvis>=2.0.0->dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Using cached pytest-3.0.7-py2.py3-none-any.whl
Collecting numexpr (from pyLDAvis>=2.0.0->dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Using cached numexpr-2.6.2-cp35-cp35m-manylinux1_x86_64.whl
Requirement already up-to-date: wheel>=0.23.0 in /mnt/data/jenkins/shiningpanda/jobs/62c67c92/virtualenvs/d41d8cd9/lib/python3.5/site-packages (from pyLDAvis>=2.0.0->dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Collecting funcy (from pyLDAvis>=2.0.0->dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Collecting pandocfilters>=1.4.1 (from nbconvert->jupyter->-r requirements-dev.txt (line 4))
Collecting bleach (from nbconvert->jupyter->-r requirements-dev.txt (line 4))
Using cached bleach-2.0.0-py2.py3-none-any.whl
Collecting entrypoints>=0.2.2 (from nbconvert->jupyter->-r requirements-dev.txt (line 4))
Using cached entrypoints-0.2.2-py2.py3-none-any.whl
Collecting nbformat (from nbconvert->jupyter->-r requirements-dev.txt (line 4))
Using cached nbformat-4.3.0-py2.py3-none-any.whl
Collecting pygments (from nbconvert->jupyter->-r requirements-dev.txt (line 4))
Using cached Pygments-2.2.0-py2.py3-none-any.whl
Collecting testpath (from nbconvert->jupyter->-r requirements-dev.txt (line 4))
Using cached testpath-0.3-py2.py3-none-any.whl
Collecting traitlets>=4.2 (from nbconvert->jupyter->-r requirements-dev.txt (line 4))
Using cached traitlets-4.3.2-py2.py3-none-any.whl
Collecting mistune!=0.6 (from nbconvert->jupyter->-r requirements-dev.txt (line 4))
Using cached mistune-0.7.4-py2.py3-none-any.whl
Collecting jupyter-core (from nbconvert->jupyter->-r requirements-dev.txt (line 4))
Using cached jupyter_core-4.3.0-py2.py3-none-any.whl
Collecting ipython>=4.0.0 (from ipywidgets->jupyter->-r requirements-dev.txt (line 4))
Using cached ipython-5.3.0-py3-none-any.whl
Collecting widgetsnbextension~=2.0.0 (from ipywidgets->jupyter->-r requirements-dev.txt (line 4))
Using cached widgetsnbextension-2.0.0-py2.py3-none-any.whl
Collecting jupyter-client (from ipykernel->jupyter->-r requirements-dev.txt (line 4))
Using cached jupyter_client-5.0.1-py2.py3-none-any.whl
Collecting tornado>=4.0 (from ipykernel->jupyter->-r requirements-dev.txt (line 4))
Collecting ipython-genutils (from qtconsole->jupyter->-r requirements-dev.txt (line 4))
Using cached ipython_genutils-0.2.0-py2.py3-none-any.whl
Collecting prompt-toolkit<2.0.0,>=1.0.0 (from jupyter-console->jupyter->-r requirements-dev.txt (line 4))
Using cached prompt_toolkit-1.0.14-py3-none-any.whl
Collecting terminado>=0.3.3; sys_platform != "win32" (from notebook->jupyter->-r requirements-dev.txt (line 4))
Collecting boto>=2.32 (from smart-open>=1.2.1->gensim>=0.13.2->dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Using cached boto-2.46.1-py2.py3-none-any.whl
Collecting bz2file (from smart-open>=1.2.1->gensim>=0.13.2->dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Collecting MarkupSafe>=0.23 (from Jinja2>=2.4->flask>=0.11.1->dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Requirement already up-to-date: setuptools in /mnt/data/jenkins/shiningpanda/jobs/62c67c92/virtualenvs/d41d8cd9/lib/python3.5/site-packages (from pytest->pyLDAvis>=2.0.0->dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Collecting py>=1.4.29 (from pytest->pyLDAvis>=2.0.0->dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Using cached py-1.4.33-py2.py3-none-any.whl
Collecting html5lib>=0.99999999 (from bleach->nbconvert->jupyter->-r requirements-dev.txt (line 4))
Using cached html5lib-0.999999999-py2.py3-none-any.whl
Collecting jsonschema!=2.5.0,>=2.4 (from nbformat->nbconvert->jupyter->-r requirements-dev.txt (line 4))
Using cached jsonschema-2.6.0-py2.py3-none-any.whl
Collecting decorator (from traitlets>=4.2->nbconvert->jupyter->-r requirements-dev.txt (line 4))
Using cached decorator-4.0.11-py2.py3-none-any.whl
Collecting pickleshare (from ipython>=4.0.0->ipywidgets->jupyter->-r requirements-dev.txt (line 4))
Using cached pickleshare-0.7.4-py2.py3-none-any.whl
Collecting simplegeneric>0.8 (from ipython>=4.0.0->ipywidgets->jupyter->-r requirements-dev.txt (line 4))
Collecting pexpect; sys_platform != "win32" (from ipython>=4.0.0->ipywidgets->jupyter->-r requirements-dev.txt (line 4))
Using cached pexpect-4.2.1-py2.py3-none-any.whl
Collecting pyzmq>=13 (from jupyter-client->ipykernel->jupyter->-r requirements-dev.txt (line 4))
Using cached pyzmq-16.0.2-cp35-cp35m-manylinux1_x86_64.whl
Collecting wcwidth (from prompt-toolkit<2.0.0,>=1.0.0->jupyter-console->jupyter->-r requirements-dev.txt (line 4))
Using cached wcwidth-0.1.7-py2.py3-none-any.whl
Collecting ptyprocess (from terminado>=0.3.3; sys_platform != "win32"->notebook->jupyter->-r requirements-dev.txt (line 4))
Using cached ptyprocess-0.5.1-py2.py3-none-any.whl
Requirement already up-to-date: appdirs>=1.4.0 in /mnt/data/jenkins/shiningpanda/jobs/62c67c92/virtualenvs/d41d8cd9/lib/python3.5/site-packages (from setuptools->pytest->pyLDAvis>=2.0.0->dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Requirement already up-to-date: packaging>=16.8 in /mnt/data/jenkins/shiningpanda/jobs/62c67c92/virtualenvs/d41d8cd9/lib/python3.5/site-packages (from setuptools->pytest->pyLDAvis>=2.0.0->dariah-topics==0.2.0.dev0->-r requirements.txt (line 1))
Collecting webencodings (from html5lib>=0.99999999->bleach->nbconvert->jupyter->-r requirements-dev.txt (line 4))
Using cached webencodings-0.5.1-py2.py3-none-any.whl
Installing collected packages: nose, coverage, nosexcover, pandocfilters, webencodings, html5lib, bleach, entrypoints, ipython-genutils, decorator, traitlets, jupyter-core, jsonschema, nbformat, pygments, testpath, MarkupSafe, Jinja2, mistune, nbconvert, pickleshare, simplegeneric, ptyprocess, pexpect, wcwidth, prompt-toolkit, ipython, python-dateutil, pyzmq, jupyter-client, tornado, ipykernel, terminado, notebook, widgetsnbextension, ipywidgets, qtconsole, jupyter-console, jupyter, pytz, numpy, pandas, regex, scipy, requests, boto, bz2file, smart-open, gensim, cycler, matplotlib, werkzeug, itsdangerous, click, flask, beautifulsoup4, wikipedia, lxml, future, joblib, py, pytest, numexpr, funcy, pyLDAvis, dariah-topics
Running setup.py develop for dariah-topics
Successfully installed Jinja2-2.9.6 MarkupSafe-1.0 beautifulsoup4-4.5.3 bleach-2.0.0 boto-2.46.1 bz2file-0.98 click-6.7 coverage-4.3.4 cycler-0.10.0 dariah-topics decorator-4.0.11 entrypoints-0.2.2 flask-0.12.1 funcy-1.7.3 future-0.16.0 gensim-2.0.0 html5lib-0.999999999 ipykernel-4.6.1 ipython-5.3.0 ipython-genutils-0.2.0 ipywidgets-6.0.0 itsdangerous-0.24 joblib-0.11 jsonschema-2.6.0 jupyter-1.0.0 jupyter-client-5.0.1 jupyter-console-5.1.0 jupyter-core-4.3.0 lxml-3.7.3 matplotlib-1.5.3 mistune-0.7.4 nbconvert-5.1.1 nbformat-4.3.0 nose-1.3.7 nosexcover-1.0.11 notebook-5.0.0 numexpr-2.6.2 numpy-1.12.1 pandas-0.19.2 pandocfilters-1.4.1 pexpect-4.2.1 pickleshare-0.7.4 prompt-toolkit-1.0.14 ptyprocess-0.5.1 py-1.4.33 pyLDAvis-2.1.1 pygments-2.2.0 pytest-3.0.7 python-dateutil-2.6.0 pytz-2017.2 pyzmq-16.0.2 qtconsole-4.3.0 regex-2017.4.5 requests-2.13.0 scipy-0.19.0 simplegeneric-0.8.1 smart-open-1.5.2 terminado-0.6 testpath-0.3 tornado-4.5 traitlets-4.3.2 wcwidth-0.1.7 webencodings-0.5.1 werkzeug-0.12.1 widgetsnbextension-2.0.0 wikipedia-1.4.0
+ ./setup.py sdist bdist_wheel
running sdist
running egg_info
writing top-level names to dariah_topics.egg-info/top_level.txt
writing requirements to dariah_topics.egg-info/requires.txt
writing dariah_topics.egg-info/PKG-INFO
writing dependency_links to dariah_topics.egg-info/dependency_links.txt
writing manifest file 'dariah_topics.egg-info/SOURCES.txt'
/mnt/data/jenkins/shiningpanda/jobs/62c67c92/virtualenvs/d41d8cd9/lib/python3.5/site-packages/setuptools/dist.py:334: UserWarning: Normalizing '0.2.0dev0' to '0.2.0.dev0'
normalized_version,
warning: sdist: standard file not found: should have one of README, README.rst, README.txt
running check
warning: check: missing required meta-data: url
creating dariah_topics-0.2.0.dev0
creating dariah_topics-0.2.0.dev0/dariah_topics
creating dariah_topics-0.2.0.dev0/dariah_topics.egg-info
creating dariah_topics-0.2.0.dev0/test
copying files to dariah_topics-0.2.0.dev0...
copying setup.cfg -> dariah_topics-0.2.0.dev0
copying setup.py -> dariah_topics-0.2.0.dev0
copying dariah_topics/__init__.py -> dariah_topics-0.2.0.dev0/dariah_topics
copying dariah_topics/doclist.py -> dariah_topics-0.2.0.dev0/dariah_topics
copying dariah_topics/evaluation.py -> dariah_topics-0.2.0.dev0/dariah_topics
copying dariah_topics/mallet.py -> dariah_topics-0.2.0.dev0/dariah_topics
copying dariah_topics/model_creation.py -> dariah_topics-0.2.0.dev0/dariah_topics
copying dariah_topics/preprocessing.py -> dariah_topics-0.2.0.dev0/dariah_topics
copying dariah_topics/visualization.py -> dariah_topics-0.2.0.dev0/dariah_topics
copying dariah_topics.egg-info/PKG-INFO -> dariah_topics-0.2.0.dev0/dariah_topics.egg-info
copying dariah_topics.egg-info/SOURCES.txt -> dariah_topics-0.2.0.dev0/dariah_topics.egg-info
copying dariah_topics.egg-info/dependency_links.txt -> dariah_topics-0.2.0.dev0/dariah_topics.egg-info
copying dariah_topics.egg-info/requires.txt -> dariah_topics-0.2.0.dev0/dariah_topics.egg-info
copying dariah_topics.egg-info/top_level.txt -> dariah_topics-0.2.0.dev0/dariah_topics.egg-info
copying test/test_fuzzy_segmenting.py -> dariah_topics-0.2.0.dev0/test
Writing dariah_topics-0.2.0.dev0/setup.cfg
Creating tar archive
removing 'dariah_topics-0.2.0.dev0' (and everything under it)
running bdist_wheel
running build
running build_py
copying dariah_topics/preprocessing.py -> build/lib/dariah_topics
installing to build/bdist.linux-x86_64/wheel
running install
running install_lib
creating build/bdist.linux-x86_64/wheel
creating build/bdist.linux-x86_64/wheel/dariah_topics
copying build/lib/dariah_topics/preprocessing.py -> build/bdist.linux-x86_64/wheel/dariah_topics
copying build/lib/dariah_topics/model_creation.py -> build/bdist.linux-x86_64/wheel/dariah_topics
copying build/lib/dariah_topics/mallet.py -> build/bdist.linux-x86_64/wheel/dariah_topics
copying build/lib/dariah_topics/evaluation.py -> build/bdist.linux-x86_64/wheel/dariah_topics
copying build/lib/dariah_topics/__init__.py -> build/bdist.linux-x86_64/wheel/dariah_topics
copying build/lib/dariah_topics/doclist.py -> build/bdist.linux-x86_64/wheel/dariah_topics
copying build/lib/dariah_topics/visualization.py -> build/bdist.linux-x86_64/wheel/dariah_topics
running install_egg_info
Copying dariah_topics.egg-info to build/bdist.linux-x86_64/wheel/dariah_topics-0.2.0.dev0-py3.5.egg-info
running install_scripts
creating build/bdist.linux-x86_64/wheel/dariah_topics-0.2.0.dev0.dist-info/WHEEL
+ nosetests
........F......F..............E...........
======================================================================
ERROR: Integration test notebook (via Jupyter)
----------------------------------------------------------------------
Traceback (most recent call last):
File "/mnt/data/jenkins/shiningpanda/jobs/62c67c92/virtualenvs/d41d8cd9/lib/python3.5/site-packages/nose/case.py", line 198, in runTest
self.test(*self.arg)
File "/mnt/data/jenkins/workspace/DARIAH-Topics/test/integration_test.py", line 20, in jupyter_integration_test
stderr=STDOUT, universal_newlines=True)
File "/usr/lib/python3.5/subprocess.py", line 316, in check_output
**kwargs).stdout
File "/usr/lib/python3.5/subprocess.py", line 398, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command '['jupyter-nbconvert', '--execute', '--log-level=ERROR', '--ExecutePreprocessor.iopub_timeout=30', '--ExecutePreprocessor.timeout=None', '/mnt/data/jenkins/workspace/DARIAH-Topics/IntegrationTest_v01.ipynb']' returned non-zero exit status 1
-------------------- >> begin captured logging << --------------------
root: ERROR: An error occurred while executing the following cell:
------------------
sparse_bow = pre.create_sparse_matrix(doc_labels, doc_tokens, id_types, doc_ids)
------------------
AttributeError: module 'dariah_topics.preprocessing' has no attribute 'create_sparse_matrix'
--------------------- >> end captured logging << ---------------------
======================================================================
FAIL: Doctest: dariah_topics.preprocessing.gensim2dataframe
----------------------------------------------------------------------
Traceback (most recent call last):
File "/usr/lib/python3.5/doctest.py", line 2190, in runTest
raise self.failureException(self.format_failure(new.getvalue()))
AssertionError: Failed doctest test for dariah_topics.preprocessing.gensim2dataframe
File "/mnt/data/jenkins/workspace/DARIAH-Topics/dariah_topics/preprocessing.py", line 755, in gensim2dataframe
----------------------------------------------------------------------
File "/mnt/data/jenkins/workspace/DARIAH-Topics/dariah_topics/preprocessing.py", line 778, in dariah_topics.preprocessing.gensim2dataframe
Failed example:
isinstance(gensim2dataframe(model), pd.DataFrame)
Exception raised:
Traceback (most recent call last):
File "/usr/lib/python3.5/doctest.py", line 1321, in __run
compileflags, 1), test.globs)
File "<doctest dariah_topics.preprocessing.gensim2dataframe[6]>", line 1, in <module>
isinstance(gensim2dataframe(model), pd.DataFrame)
File "/mnt/data/jenkins/workspace/DARIAH-Topics/dariah_topics/preprocessing.py", line 788, in gensim2dataframe
topics_df.loc[idx] = temp
File "/mnt/data/jenkins/shiningpanda/jobs/62c67c92/virtualenvs/d41d8cd9/lib/python3.5/site-packages/pandas/core/indexing.py", line 141, in __setitem__
self._setitem_with_indexer(indexer, value)
File "/mnt/data/jenkins/shiningpanda/jobs/62c67c92/virtualenvs/d41d8cd9/lib/python3.5/site-packages/pandas/core/indexing.py", line 579, in _setitem_with_indexer
value=value)
File "/mnt/data/jenkins/shiningpanda/jobs/62c67c92/virtualenvs/d41d8cd9/lib/python3.5/site-packages/pandas/core/internals.py", line 3168, in setitem
return self.apply('setitem', **kwargs)
File "/mnt/data/jenkins/shiningpanda/jobs/62c67c92/virtualenvs/d41d8cd9/lib/python3.5/site-packages/pandas/core/internals.py", line 3056, in apply
applied = getattr(b, f)(**kwargs)
File "/mnt/data/jenkins/shiningpanda/jobs/62c67c92/virtualenvs/d41d8cd9/lib/python3.5/site-packages/pandas/core/internals.py", line 740, in setitem
values[indexer] = value
ValueError: cannot copy sequence with size 4 to array axis with dimension 10
-------------------- >> begin captured logging << --------------------
gensim.corpora.dictionary: INFO: adding document #0 to Dictionary(0 unique tokens: [])
gensim.corpora.dictionary: INFO: built Dictionary(4 unique tokens: ['testing', 'corpus', 'for', 'test']) from 2 documents (total 4 corpus positions)
gensim.models.ldamodel: INFO: using symmetric alpha at 1.0
gensim.models.ldamodel: INFO: using symmetric eta at 0.25
gensim.models.ldamodel: INFO: using serial LDA version on this node
gensim.models.ldamodel: INFO: running online LDA training, 1 topics, 1 passes over the supplied corpus of 2 documents, updating model once every 2 documents, evaluating perplexity every 2 documents, iterating 1x with a convergence threshold of 0.001000
gensim.models.ldamodel: WARNING: too few updates, training might not converge; consider increasing the number of passes or iterations to improve accuracy
gensim.models.ldamodel: DEBUG: bound: at document #0
gensim.models.ldamodel: INFO: -1.685 per-word bound, 3.2 perplexity estimate based on a held-out corpus of 2 documents with 4 words
gensim.models.ldamodel: INFO: PROGRESS: pass 0, at document #2/2
gensim.models.ldamodel: DEBUG: performing inference on a chunk of 2 documents
gensim.models.ldamodel: DEBUG: 0/2 documents converged within 1 iterations
gensim.models.ldamodel: DEBUG: updating topics
gensim.models.ldamodel: INFO: topic #0 (1.000): 0.250*"test" + 0.250*"corpus" + 0.250*"testing" + 0.250*"for"
gensim.models.ldamodel: INFO: topic diff=0.217711, rho=1.000000
--------------------- >> end captured logging << ---------------------
======================================================================
FAIL: Doctest: dariah_topics.preprocessing.save_sparse_bow
----------------------------------------------------------------------
Traceback (most recent call last):
File "/usr/lib/python3.5/doctest.py", line 2190, in runTest
raise self.failureException(self.format_failure(new.getvalue()))
AssertionError: Failed doctest test for dariah_topics.preprocessing.save_sparse_bow
File "/mnt/data/jenkins/workspace/DARIAH-Topics/dariah_topics/preprocessing.py", line 560, in save_sparse_bow
----------------------------------------------------------------------
File "/mnt/data/jenkins/workspace/DARIAH-Topics/dariah_topics/preprocessing.py", line 580, in dariah_topics.preprocessing.save_sparse_bow
Failed example:
save_sparse_bow(sparse_bow, 'sparsebow')
Exception raised:
Traceback (most recent call last):
File "/usr/lib/python3.5/doctest.py", line 1321, in __run
compileflags, 1), test.globs)
File "<doctest dariah_topics.preprocessing.save_sparse_bow[5]>", line 1, in <module>
save_sparse_bow(sparse_bow, 'sparsebow')
File "/mnt/data/jenkins/workspace/DARIAH-Topics/dariah_topics/preprocessing.py", line 592, in save_sparse_bow
with open('.'.join([output_path, 'mm']), 'w', encoding="utf-8") as f:
NameError: name 'output_path' is not defined
----------------------------------------------------------------------
File "/mnt/data/jenkins/workspace/DARIAH-Topics/dariah_topics/preprocessing.py", line 582, in dariah_topics.preprocessing.save_sparse_bow
Failed example:
os.path.isfile('sparsebow.mm')
Expected:
True
Got:
False
Name Stmts Miss Cover
-----------------------------------------------------
dariah_topics.py 0 0 100%
dariah_topics/doclist.py 89 14 84%
dariah_topics/evaluation.py 100 80 20%
dariah_topics/mallet.py 216 173 20%
dariah_topics/model_creation.py 60 43 28%
dariah_topics/preprocessing.py 202 9 96%
dariah_topics/visualization.py 165 136 18%
-----------------------------------------------------
TOTAL 832 455 45%
----------------------------------------------------------------------
Ran 42 tests in 4.883s
FAILED (errors=1, failures=2)
Build step 'Virtualenv Builder' marked build as failure
Recording test results
Skipping Cobertura coverage report as build was not UNSTABLE or better ...
[Set GitHub commit status (universal)] ERROR on repos [GHRepository@7b357345[description=A python library for topic modeling.,homepage=,name=Topics,license=<null>,fork=true,size=96681,milestones={},language=Jupyter Notebook,commits={},url=https://api.github.com/repos/DARIAH-DE/Topics,id=69341969]] (sha:b193131) with context:DARIAH-Topics
Setting commit status on GitHub for https://github.com/DARIAH-DE/Topics/commit/b193131a17023fa0cc9b8d0795f2399dd77d48fe
[BFA] Scanning build for known causes...
[BFA] No failure causes found
[BFA] Done. 0s
Started calculate disk usage of build
Finished Calculation of disk usage of build in 0 seconds
Started calculate disk usage of workspace
Finished Calculation of disk usage of workspace in 0 seconds
Notifying upstream projects of job completion
Finished: FAILURE