How to view all the default HHVM options and settings? - hhvm

In this question at HHVM poor performance it mentions changing the JitWarmupRequests value for HHVM. But is there a command or way to see all default set values for all HHVM options ?
Similar to how mysql server can see all set values via mysqladmin variables command ?

No there is no way to do this in HHVM. I'm working on it, but nothing yet.

Related

Parametrize Connections from Database parametrization table

We are using Pentaho Data Integration V7 working with multiple data origins with an Oracle DWH destiny.
We have stored all the connection access data in a parametrization table, let's call it : D_PARAM. All the connections are configured using parameters (${database_name} ... etc)
We have , at the begining of every job , a transformation with a "set variables" step which reads the right parameters from D_PARAM.
This all works fine, my problem is :
Every time we want to edit a single transformation, or in the development process of a new one , we can't use the paremetrized connections because the parameters haven't been setted. We need then to use "hardcoded" connections during the development process.
Is there a better way to manage this situation ? The idea of having the connections parametrized is to avoid errors and simplify the connections management, but if at the end we need both kind of connections.. I don't see them so useful.
There's not a simple answer, you could rotate your kettle.properties file to change default values, you keep all the values in the file:
D_PARAM = DBN
D_PARAM_DB1 = DB1
D_PARAM_DB2 = DB2
...
And just update the D_PARAM with the one you need from the different D_PARAM_DBN before starting PDI. It's a hassle to be constantly updating the kettle.properties file, but works out of the box.
You could also try working with environments, for this you would have to install a plugin available in Github: https://github.com/mattcasters/kettle-environment, it was created by a former PDI developer, and I don't know if it works with v7 version, it was updated to work with 8.2, but it would probably work with v7, to test it, you can install your PDI version on another directory on your PC and install there the plugin (and other additional plugins you have in your current installation), so you don't break your setup. This blog entry gives you details on how to use the environments: http://diethardsteiner.github.io/pdi/2018/12/16/Kettle-Environment.html
I don't know if the environments plugin would solve your problem, because you can't change the environment in the middle of a job, but for me, with the maitre script to use the environments when I program a job or transform, it's been easier to work with different projects/paths in my setup.
In Spoon you can click on the “Edit” menu and “Set environment variables”. It’ll list all variables currently in use and you can set their values. Then the transformation will use those values when you run.
Also works in Preview, but it’s somewhat buggy, it doesn’t always take updated values.

Terraform - self.default_ip_address / ipv4_address missmatch

I have some annoying problem here.
I am trying to use file and remote-exec provisioners when creating (Windows) VM-s with Terraform.
Please see my template (this is only for a VM, I separated it from the variables and the core infrastructure for now) here: https://code.prkr.li/60463d5e2b38ad032b4b6678
The only thing that does not work presently is using the self.default_ip_address as a host for the provisioner's connection. The task (as seen in the Terraform logs in Jenkins) tries to connect to the IP address that the VM got from DHCP...
...instead of what is defined in the (customization) config...
So I suppose viable solutions would be:
refreshing the default IP
using hostname (not sure if that is possible, did not manage to find the right syntax)
???
Any ideas / suggestion please?
Not that there were a lot of answers, but I found my solution apparently.
Anyone having similar issue, suggested to use
host = self.clone.0.customize.0.network_interface.0.ipv4_address
Workes for me anyway.

How to modify path where Torch Hub models are downloaded

When I download models through Torch Hub, models are automatically downloaded in /home/me/.cache/torch.
How can I modify this behavior ?
From official documentation, there is several ways to modify this path.
In priority order :
Calling hub.set_dir()
$TORCH_HOME/hub, if environment variable TORCH_HOME is set.
$XDG_CACHE_HOME/torch/hub, if environment variable XDG_CACHE_HOME is set.
~/.cache/torch/hub
So I just had to do :
export TORCH_HUB=/my/path/
Edit
TORCH_HUB appear to be deprecated, use TORCH_HOME instead
I'm sorry but I'm trying to find out how to free up space, and these answers don't make sense to me.
Where do you type "TORCH_HOME"?
python TORCH_HOME in cmd brings up an error in windows 10.
and Torch and TORCH_HOME is not recognized as an internal or external command,
operable program or batch file.
Are you answering this for linux? because cmd uses a > and not $

Permanent environment variables

I have a problem assessing my advanced settings of my computer, anytime i get to the properties of my computer it freezes and vanishes, I decided to set my path through the command prompt. Though the lines below works. Its just temporal and the path is unset immediately i close the running command prompt,
set path=%path%;C:\python27
I am just wondering why my computer's property panel vanishes.
Any others ways to set the environment variables permanent?
try using setx PATH "%PATH%;C:\python27"
Note that u need to close the current cmd window, after executing this command as the changes will reflect in newer instances only.

Virtualized Puppet master showing alphanumeric error when trying to load the dashboard

When I try to go to the puppet dashboard it loads up a white page saying
The environment must be purely alphanumeric, not ''
I googled this error a bit but there does not seem to be a finite solution yet. As fast as what I've done I've looked into the conf.d files, but I'm afraid to make the wrong edit. I'm still new to this and was wondering if anyone has ever ran into a similar situation?
This happens if one tries to access puppet console in wrong port (8140) instead of 443, assuming default ports are being used.
This also happens when you install puppet-passenger The opensource.
In this case - everything is OK!
Install Open Source agent and be a Puppet Master :)

Resources