Power BI Embedded, updating a dashboard - asp.net-mvc

I am referring to the problem of updating a dashboard that has been published to azure via power BI embedded (Azure tutorial).
When I publish to azure an updated version of a dashboard (using step 6 of the desktop solution provided on the link above), I am able to publish the updated pbix file to the same workspace and dataset name. When I retrieve the list of datasets for the workspace I can see both the old and new dashboard with the same name and different id. I have found this confusing as I would have expected using the same name would overwrite the old version of the pbix file.
What would be the recommended procedure to update a dashboard? Would it be to use a new name for the dataset each time? This doesn't seem ideal as it can have implications as well in the embedding web app.

From my own experience what I have done is to just delete the original and upload again with the same information. I don't rely on the IDs generated from the power bi workspace, I keep track of everything in a table I control. For example I have a table which holds the power bi meta data with an ID I give it. If I ever need to upload a new version of the PBIX I would delete the one from the azure workspace and upload a new one using the same information and then get that new ID and store it with my local ID.
So in use I would look up the report based on my local ID in my application to get the information needed to pass to power bi api to then view the report.
Hope this helps.

Related

Create jira ticket in logic apps. Different jira project depending on environment

I can use the JIRA connector with the Create a new issue (V2) action to create a jira ticket using logic apps. However, I want to put the project name in a parameter file as it will vary depending on whether the logic app is in dev or production.
The top part of the screenshot is my attempt at using a variable which comes from a parameter file. The bottom part is when I hard code the project name.
How can I vary the name of the JIRA project depending on whether it's in dev or production using the JIRA connector in logic apps?

save/load thingsboard configuration

Is it possible to somehow serialize current Thingsboard (let's call it TBoard) configuration, save it and than latter load saved configuration on TBoard startup.
I am specifically interested in loading device profiles, rule chains, and dashboards.
I want to save configuration together with my project in git repository so than latter I could just use docker-compose to start multiple services from project (let's call them sensors) and single TBoard instance with saved configuration which will be used for collecting telemetry from sensors and drawing dashboards.
Another reason for saving configuration is what happens if for some reason TBoard container crashes or somehow get corrupted so it can't be started again, would I have to click on the things again in order to create all device profiles, dashboards, configure rule chains ... etc etc ... ?
Regarding this line
I am specifically interested in loading device profiles, rule chains, and dashboards. I want to save configuration together with my project in git repository
I have just recently implemented version control for my Thingsboard deployment. The way i am doing it is with the python REST client.
I have written functions to export all dashboards/data converters/integrations/rule chains/widgets into json files which I save into a github repository.
I have also written the reverse script to push the stored files to a fresh environment, essentially "flashing" it. Surprisingly, this works perfectly.
I have an idea to publish this as a package, but it's something I've never done before so I'm unsure if I will get to it.
Just letting you know that it is definitely possible to get source control operational via the API.

Azure DevOps secure file guids

In my ADO build pipline, I have a secure file download step. When we branch versions, we use powershell to do the heavy lifting with cloning build definitions and updating settings/info in the cloned pipeline.
One issue I've run into is that the Secure File Download step doesn't accept variables, and in the UI you can only select names of files that already exist, so we've had to manually update it after every new branch we create.
I've grabbed the definition task step in powershell (as $step) and was hoping I could set the $step.inputs.fileInputs to a variable I assign to something like cert-$newVersion, however it currently is set to a guid.
Does anyone know if it possible to get the guid of secure files in ADO via the API or have a solution?
Does anyone know if it possible to get the guid of secure files in ADO via the API or have a solution?
Yes. This API exists.
You could try to use the following Rest API:
Get https://dev.azure.com/{OrganizationName}/{ProjectName}/_apis/distributedtask/securefiles?api-version=6.1-preview.1
Result:
You could get the secure file GUID based on the file name.

WSO2 loss APIs after changes in docker container

I'm having another problem using WSO2 API Manager 2.0.0: I have installed it in docker using three containers (one for APIM, one for Analytics and one for MySQL) and I replace some configuration files with my custom version (e.g. DB, server name, gateway setup...).
Both APIM and Analytics are configured to save data in the MySQL container and I am able to see changes in the DB.
The issue is that I cannot find my APIs neither in the publisher nor in the store after the container has been rebuilt. Changes in the DB persists, I can see the statistics for all my APIs and I get an error if I try to create a new API using the same name or context, but the store is always empty after a new build.
I have also tried to put both /repository/deployment/server/synapse-config/default and /repository/tenants/ in two volumes and I can see the files created in /.../default/api/ for my APIs, but I cannot figure out the issue.
Should I persists some additional directory not mentioned in the guide?
I don't want to put the whole APIM and Analytics homes in volumes if possible.
First, check whether artifacts can be located in Resources Browser.
If you can find the API related files, then the issue is related to indexing.
Do the following to re-index the artifacts in the registry:
Rename the <lastAccessTimeLocation> element in the <APIM_2.0.0_HOME>/repository/conf/registry.xml file. If you use a clustered/distributed API Manager setup, change the file in the API Publisher node. For example, change the /_system/local/repository/components/org.wso2.carbon.registry/indexing/lastaccesstime registry path to /_system/local/repository/components/org.wso2.carbon.registry/indexing/lastaccesstime_1.
Shut down the API Manager, back up and delete the <APIM_2.0.0_HOME>/solr directory.
Finally start the API Manager.
The Api Information resides in the DB and in the File system.(/repository/deployment/server/synapse-config/default/api) It is possible that the registry artifacts are not indexed properly. Can you try following?
Delete the solar directory.
Open registry.xml and change the following line as shown below. < lastAccessTimeLocation>/_system/local/repository/components/org.wso2.carbon.registry/indexing/lastaccesstime-1
Now restart the server. Server will re-index all the files again.
Also make sure the Databases are properly configured. Specially Registry mounting related configurations.

Where are TFS Admin Console Settings Saved?

this may sound like a silly question but our IT department wants us to use their approved backup tool instead of using the standard back solution that comes packaged within TFS2010. My question is, is there a config file holding all settings found within the TFS Admin Console or are all these values saved within the TFS_Configuration table.
Please note that we are using TFS2010
Thanks
The settings are indeed stored in the config database.

Resources