Jenkins sphinx-build not found - jenkins

I'm using default sphinx-quickstart and Ubuntu 18.01
I'm having this issue with Jenkins when I try to do make html
/bin/sh: 1: sphinx-build: not found
Makefile:19: recipe for target 'html' failed
It works fine when I do the make html while in a terminal...
What could be the issue?
More information: I'm using the same user (tested with whoami) and I have installed sphinx using pip

I found a workaround based on one answer here
I changed the SPHINXBUILD to SPHINXBUILD = python -m sphinx. I'm not sure this is the best way to do it, but at least it works.
I'm pretty sure my original bug has something to do with make and sh executing on its own context and not finding sphinx-build. If someone has a better answer, please post :)

Related

running pycharm interpreter using nvidia-docker2

Im working on Ubuntu 20. I've installed docker, nvidia-docker2. On Pycharm, I've followed jetbrain guide, but in the advanced steps it isn't consistent with what I see in my setup. I use PyCharm Proffesional 2022.2.
In this step:
in the run options I put additionally --runtime=nvidia and --gpus=all.
Step 4 finishes as same as in the guide (almost, but it seems that it doesn't bother anything so on that later) and on step 5 I put manually the path to the interpreter in the virtual environment I've created using the Dockerfile.
In that way I am able to run the command of nvidia-smi and see correctly the GPU, but I don't see any packages I've installed during the Dockerfile build.
There is another option to connect the interpreter a little bit differently in which I do see the packages, but I can't run the nvidia-smi command and the torch.cuda.is_availble return False.
The way is instead of doing this as in the guide:
I press on the little down arrow in left of the Add Interpreter button and then click on Show all:
After which I can press the + button :
works, so it might be PyCharm "Python Console" issue.
and then I can choose Docker:
which will result in the difference mentioned above in functionality and also in the path dispalyed (the first one is the first remote interpreter top to bottom direction and the second is the second correspondingly):
Here of course the effect of the first and the second correspondingly:
Here is the results of the interpreter run with the first method connected interpreter:
and here is the second:
Of the following code:
Here is the Dockerfile file if you want to take a look:
Anyone configured it correctly and can help ?
Thank you in advance.
P.S: if I run the docker from services and enter the terminal the command nvidia-smi works fine and also the import of torch and the command torch.cuda.is_available return True.
P.S.2:
The thing that has worked for me for now is to change the Dockerfile to install directly torch with pip without create conda environement.
Then I set the path to the python2.7 and I can run the code, but not debug it.
for run the result is as expected (the packages list as was shown before is still empty, but it works, I guess somehow my IDE cannot access the packages list of the remote interpreter in that case, I dont know why):
But the debugger outputs the following error:
Any suggestions for the debugger issue also will be welcome, although it is a different issue.
Please update to 2022.2.1 as it looks like a known regression that has been fixed.
Let me know if it still does not work well.

Sawtooth network and CoopEdge - installation problem with dockerfile

I'm learning about blockchain and now I'm starting with Sawtooth network as I have heard that it's quite popular. I came across and research paper about CoopEdge and it's very interesting (github link: https://github.com/coopedge/prototype).
However, I don't know clearly how to make this work. I personally send an email and still don't have the response so I have to try it with basic knowledge (blindly, somehow). There are two folder sawtooth-core and sawtooth-poer. I went with poer because the publication was talking about it. There are two type of dockerfile - docker-compose.yml and docker-compose-installed.yaml. I installed the first one with docker-compose and there was no problem. However, when I tried to install the latter one I keep getting the error:
Step 9/13 : RUN export VERSION=$(./bin/get_version) && sed -i -e "0,/version.*$/ s/version.*$/version\ =\ \"${VERSION}\"/" Cargo.toml && /root/.cargo/bin/cargo deb --deb-version $VERSION
---> Running in 7d244fc29e30
/bin/sh: 1: ./bin/get_version: Permission denied
91mcargo-deb: Argument to option 'deb-version' missing
I tried several methods by searching the internet but no luck so far. I also tried to install with root (sudo -i) but still it doesn't work at all.
Another thing is I don't know the second docker-file is mandatory for installation as there is no document or guidance provided by the author.
I appreciate any help that could solve this permission problem. Thank so much.
P.s: I'm using virtual machine with Ubuntu 18.04.

Airflow on windows 10 - Module not found errors

I'm new to data science and wanted to do a little tutorial, which requires airflow, among other things. I installed it on windows using git bash in VS Code. I tried running it but it had a problem not being able to load the sqlite3 import
command (module not found error). I figured out that if I added the directory of sqlite3.py to the path, it would run, but now it gives me a similar error: pwd module not found from daemon.py
File "C:\ProgramData\Anaconda3\lib\site-packages\daemon\daemon.py", line 18, in <module>
import pwd
ModuleNotFoundError: No module named 'pwd'
Strange to me that it can't find pwd. Obviously pwd works in both git bash and powershell natively. It seems like a basic universal command. I'd love to learn more about what's going on. I don't want to have to end up adding 100 things to path just to get this program to run. I'd love any insights anyone can provide.
PS I'm using Anaconda.
it's seems to be the side effects of Spawning new Python Daemons .
You likely can fix this by downgrading the Python-Daemon :
pip install python-daemon==2.1.2

launcher.gcr.io/google/jenkins2 does not draw chart

I am using launcher.gcr.io/google/jenkins2 to run jenkins in gcp kubernetes engine.
Everything seems ok except that I get Could not initialize class org.jfree.chart.JFreeChart error for every chart that jenkins attempt to draw. I googled the error and almost everyone solves that with adding -Djava.awt.headless=true. As you can guess I already tried that and it does not work.
Ideas?
One other possible solution/workaround is seen in JENKINS issue 39636:
I installed libxext6 libxrender1 fontconfig libfontconfig but it didn't help. After that I also installed libjfreechart-java but I still have the same problem.
Commenting out the assistive_technologies line in /etc/java-8-openjdk/accessibility.properties solved it.
You can see that recommendation in tianon's comment of that answer:
In my case it ended up being bugs.debian.org/798794 (from "Assistive technology not found error while building aprof-plot").
Adding "RUN sed -i 's/^assistive_technologies=/#&/' /etc/java-8-openjdk/accessibility.properties" to my Dockerfile fixed it. :)

PsExec is not recognized as an internal or external command

I have a job that needs to run a script on a remote computer. I'm doing so by using psexec via "Execute windows batch command":
C:\PsExec.exe \\computername -u username -p password -accepteula c:\xxx.exe
When I run the job I get the following error:
c:\PsExec.exe is not recognized as an internal or external command
** PsExec.exe is located under c:\
Any ideas?
First Define psexec.exe path in environment varaiable "PATH" or else place psexec.exe file in C:\Windows\System32\
And to Download Psexec.exe file
https://download.sysinternals.com/files/PSTools.zip
One possible explanation is the version of PsExec.exe: 32bits or 64bits.
If you have the 32 one on a 64bits machine, that command would not be recognized indeed. PsExec64.exe would.
I can see the age of this question and my answer may not be relevant to this topic since I was technically trying to solve a different problem, but maybe this will help other people who are stuck.
c:\PsExec.exe is not recognized as an internal or external command
I was trying to disable the Maintenance Configurator with PSExec (my problem is the never ending maintenance bug) and kept running into the same error as the OP BUT I got PSexec64 to run this command:
C:\PsExec64.exe -s schtasks /change /tn >"\Microsoft\Windows\TaskScheduler\Maintenance Configurator" /DISABLE
BY checking the "Run this program as an administrator" option under the Compatibility settings for "PsExec64.exe"
Don't know if this has solved my problem yet, but I think the OP would have been able to run his process if he had done this. Dear OP did you ever solve that?

Resources