Python 3.5 searching for libpython3.6m.so.1.0 - docker

I have a docker image based from python:3.5-slim, that I added Oracle 12 on it, so my tests on GitLab can run in it.
When GitLab start building the environment, like doing the pip install, I get this error message:
"/builds/root/ppc_busca/venv/bin/python3: error while loading shared libraries: libpython3.6m.so.1.0: cannot open shared object file: No such file or directory"
The thing is, how can it complains about the file 'libpython3.6m' if my python version is 3.5? That is something that I don't understand.
Thanks for the help!

Related

Problem understanding how to, if at all possible, run my docker file (.tar)

I received a .tar docker file from a friend that told me that it should contain all dependences for a program that I've been struggling to get working and that all I need to do is "run" the Docker file. The Docker file is of a .tar format and is around 3.1 GB. The program this file was setup to run is call opensimrt. The GitHub link to the file is as follows:
https://github.com/mitkof6/OpenSimRT
The google drive link to the Docker file is as follows:
https://drive.google.com/file/d/1M-5RnnBKGzaoSB4MCktzsceU4tWCCr3j/view?usp=sharing
This program has many dependencies, some big ones to note is that it runs off ubuntu 18.04 and Opensim 4.1.
I'm not a computer scientist by any means, so I've been struggling to even learn to do docker basics like load and run a image. However, I desperately need this program to work. If you have any steps or advice on how to run this .tar I'd greatly appreciate it. Alternatively if you are able to find a way to get opensimrt up and running and can post those steps I'd be more than happy with that solution as well.
I've tried the commands "docker run" and "docker load" followed by their respective tags, file paths, args..etc. However, even when I fix various issues I always get stuck with a missing var/lib/docker/tmp/docker-import-....(random numbers) file. The numbers change every so often when trying to solve the issue, but eventually I always end up getting some variation of this error: Error response from daemon: open /var/lib/docker/tmp/docker-import-3640220538/bin/json: no such file or directory.
ps: I have extracted the .tar already and there is no install guide/instruction, .exe, install application. As a result I'm not sure how to get the program installed and running.

python bundled by py2app still points to local path

I would like to create a macOS app using py2app. My application (testapp) has a few dependencies
(paramiko,scp,PyYAML,PyQt5). To do so:
I installed python3 using brew (located in /usr/local/Cellar).
I installed the aforementioned dependencies in this python
I ran python3 setup.py py2app --packages paramiko,yaml,scp
Everything runs apparently fine. I have a testapp.app directory created in dist. However, when I run the python located in the bundle (./dist/testapp.app/Contents/MacOS/python) I run into troubles. Indeed, I can import the aforementioned external packages but when I inspect their path, the result gives me the local installation and not the one of the bundle.
For instance paramiko.__path__ gives me: /usr/local/Cellar/python3/3.6.3/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/paramiko
Looking deeper into the bundle, I could not find the file site-packages.zip which should be created by py2app.
As such my bundle is absolutely not self consistent and can not be distributed on other machines. Would you have any idea about what is wrong with my setup ?
EDIT: I forgot to mention that I started to inspect the embedded python because running the embedded app failed.

missing libary in vitis-tool-gpu docker container

I've used the Vitis docker tool container using only the CPU and Conda worked fine; however, when I want to use the GPU version for docker, I get the below error. I tried building the environment twice and each time it has failed to import the right libraries.
(vitis-ai-caffe) sam#Itec:~/cf_resnet50$ vai_q_caffe quantize -model float/trainval.prototxt -weights float/trainval.caffemodel
vai_q_caffe: error while loading shared libraries: libprotobuf.so.21: cannot open shared object file: No such file or director
How might I resolve this error? I will be happy to receive any help.
With version 1.0.0 there is an issue in the docker with the protobuf library version. It can be resolved by patching the conda_requirements.txt. To do so, open the conda_requirements.txt in a text editor and add the following two lines, then build the GPU docker:
libprotobuf==3.10.1
protobuf==3.10.1

Bazel internal shell issue using windows

I am trying to migrate a huge project having visual studio and maven projects to bazel. I need to access our in house maven server which is encrypted. To get access I need the load the maven_jar skylark extension since the default impl does not support encryption (get error 401). using the extension leads to a lot of troubles, like:
ERROR: BUILD:4:1: no such package '#org_bouncycastle_bcpkix_jdk15on//jar': Traceback (most recent call last):
File ".../external/bazel_tools/tools/build_defs/repo/maven_rules.bzl", line 280
_maven_artifact_impl(ctx, "jar", _maven_jar_build_file_te...)
File ".../external/bazel_tools/tools/build_defs/repo/maven_rules.bzl", line 248, in _maven_artifact_impl
fail(("%s: Failed to create dirs in e...))
org_bouncycastle_bcpkix_jdk15on: Failed to create dirs in execution root.
The main issue seems to be the shell that needs to be provided to bazel in BAZEL_SH environment variables:
I am working under windows
I am using bazel 0.23.2
bazel seems to run a bash command using "bash" directly and not the one provided by env variable.
I got a ubuntu shell installed in windows. bazel was using everything from ubuntu, especially when using maven (settings.xml was using from ubuntu ~/.m2 and not from windows user)
after uninstalling ubuntu and making sure that bash in a cmd ends up in "command not found" I also removed the BAZEL_SH env var and bazel throws the message above
after setting the BAZEL_SH variable again it fails with the same error message
I am assuming that bazel gets a bash from somewhere or is ignoring the env variable. My questions are:
1. How to setup a correct shell?
2. Is BAZEL_SH needed when using current version?
For me the doc at bazel website about setup is outdated.
Cheers
Please consider using rules_jvm_external to manage your Maven dependencies. It supports both Windows and private repositories using HTTP Basic Authentication.
For me the doc at bazel website about setup is outdated.
The Bazel team is aware of this and will be updating our docs shortly.

Cryptogen tool not found. Hyperledger fabric installation in Windows 10

I followed the Hyperledger fabric documentation to install and configure it in Windows 10. However when I run the command - "./byfn.sh -m generate" for first-network sample application, I get the following error,
I have gone thru all StackOverflow questions regarding this and made sure following steps are done,
Have set the $PATH variable correctly to include bin folder.
Have downloaded the platform-specific binary and my bin folder looks like this,
I have doubts about following steps,
I have installed Docker for Windows and was able to verify the docker installation by running hello-world image in Docker. However, I have not shared any of my local drives in Docker. Not sure whether this is the cause of this error.
Please note that this is my first question in StackOverflow. Forgive me for any mistakes/redundancies. Any help is greatly appreciated.
I'd suggest making sure that you run the script to download / install the binaries and images from within the fabric-samples directory.
The $Path is exported every time you run the byfn.sh script, confirm that the path configuration in the byfn.sh is correct and points to your correct bin location
# prepending $PWD/../bin to PATH to ensure we are picking up the correct binaries
# this may be commented out to resolve installed version of tools if desired
export PATH=${PWD}/../../bin:${PWD}:$PATH
export FABRIC_CFG_PATH=${PWD}

Resources