Deploy OVF to ESXI with 2 vmdk files using pyvmomi python librafy - esxi

I want to deploy an OVF using the https://github.com/vmware/pyvmomi-community-samples/blob/master/samples/deploy_ovf.py . This utility provide way when only one vmdk file. Anybody knows how to make this work with and OVF with 2 vmdk files

Related

How can I package a windows dependent program into a lightweight docker container?

I currently have a folder containing some .dll files, .bin files and some .exe files. The main .exe that I will be executing only works on windows, and I am not entirely sure what are all its dependencies. My goal is to package all the files in the folder into a docker container so I can integrate it into the rest of my pipeline. The main .exe is a command line tool which is only called once with some arguments and left to run.
I have tried to use windows server core as the container image and it works. However this image is too big for my needs. I have tried to use nano server but when I try to run the executable there is nothing printed in the command line and the program does not run. In that scenario, if I type:
C:\Bin\x64>echo %ERRORLEVEL%
I get the following output:
-1073741515
Meaning I am missing some dependencies.
So, I'm wondering if there is an alternate solution to packaging this folder since windows server core is too big.
Most likely you'll have to stick with the Server Core image. The main thing is that these images server different purposes, and the Nano Server is intended to new applications being developed with the targeted Nano Server API. Server Core is the image focused on existing apps, but its APIs make the image larger than what one would expect from a container.
Keep in mind that it's still better than a full VM. :)
I blogged about this here: https://techcommunity.microsoft.com/t5/containers/nano-server-x-server-core-x-server-which-base-image-is-the-right/ba-p/2835785

want to run docker file in windows machine

Team mate sent me docker file - zipped one. I want to run it in my machine. I am using windows 10 64bit machine. Do I need to unzipped file first? I am new to docker. Please suggest
Thanks
Yes, you need to unzip the Dockerfile and place it in a folder. You also need to setup Docker For Windows which can be found here: https://docs.docker.com/v17.09/docker-for-windows/install/
After setting up, start docker and use
docker build command
in the folder where you placed the Dockerfile to build the image.

Deploy features.xml in servicemix during jenkins Build

I have my features.xml file in src/main/resources/features folder , when I build my project through Jenkins after building my bundle goes to the nexus repository , my requirement is that after my bundle goes to nexus then features.xml should automatically be deployed on servicemix as part of build only. I should not open the servicemix console to install the feature. Please help
You may think about using a KAR (KAraf aRchive).
More information can be found here: http://karaf.apache.org/manual/latest-3.0.x/users-guide/kar.html
You can build а KAR (through Jenkins), containing your feature, then you can use a hot deployment.
Apache Karaf also provides a KAR deployer. It means that you can drop
a KAR file directly in the deploy folder.
Apache Karaf will automatically install KAR files from the deploy
folder. You can change the behaviours of the KAR deployer in the
etc/org.apache.karaf.kar.cfg:
I have also been working on this and my solution was to turn to automated scripting to accomplish this. I wrote a ssh and FTP based program which would stop an smx, delete the ${karaf.home}/data/cache/ directory, replace the new feature file with the one retrieved from the ftp operation, then restart the karaf container.
If you are open to looking into other possibilities:
You can look into Fuse Fabric which can link many smx Containers together and implement version increases and rollbacks. Currently I believe this would also need scripting to accomplish it automatically.
The third option is relatively new and comes in the form of Building docker images and deploying them via OpenShiftV3 which was just unveiled at the Redhat Summit 2015. Its worth noting its fairly new, but it does pack a very impressive feature set.

How to create folder and upload file inside openshift?

I'm using rhc client tools for openshift. and i want to upload it PhantomJS. So, I've Executable file and now i want to upload into my openshift server, So please let me know. How can i create folder? and how to upload it? I googled but fully confused with ${OPENSHIFT_DATA_DIR}.
Help would be appreciated!
If you are on Windows you can use winscp to do what you want. Use the equivalent for Mac or Linux. You will want to upload to the DATA directory in your gear. The data directory is already created for you.

xe6 - auto deploy output to AWS S3 using REST

OK, new to rad studio, and enjoying it.
We have a cross=platform product, and would like to configure XE6 to automatically deploy to compiled .exe & Dll files to specified buckets in S3.
We have looked at the cloud exploerer example which came with samples and can load the files manually using REST.
http://docwiki.embarcadero.com/CodeExamples/XE6/en/DataSnap.Cloud_Explorer_Sample
and have checked out how to specify an output folder
Where can I specify an output folder for DLLs?
As discussed we are not that familiar with XE6, and while they have something called a 'platform deployment manager' we don't want to setup up and EC2 instance to facilitate the process.
Does anyone have any sample scripts or advice on how to configure the output path to use the REST protocol to put the compiled files up on S3 automatically?

Resources