I deployed spring-cloud-dataflow-server-cloudfoundry to SAP Cloud Fondry with environments below:
SPRING_CLOUD_DEPLOYER_CLOUDFOUNDRY_URL:https://api.cf.sap.hana.ondemand.com
SPRING_CLOUD_DEPLOYER_CLOUDFOUNDRY_ORG:{org}
SPRING_CLOUD_DEPLOYER_CLOUDFOUNDRY_SPACE:{space}
SPRING_CLOUD_DEPLOYER_CLOUDFOUNDRY_DOMAIN:{doamin}
SPRING_CLOUD_DEPLOYER_CLOUDFOUNDRY_USERNAME:username
SPRING_CLOUD_DEPLOYER_CLOUDFOUNDRY_PASSWORD:password
SPRING_CLOUD_DEPLOYER_CLOUDFOUNDRY_SKIP_SSL_VALIDATION:false
SPRING_CLOUD_DEPLOYER_CLOUDFOUNDRY_STREAM_SERVICES: mq
And import stream starter apps using bulk import applications.
And I create stream using "time-source-rabbit-1.3.0.RELEASE.jar" and "log-sink-rabbit-1.3.0.RELEASE.jar".
But I cannot deploy stream.
The status is "partial" fianlly, and apps' runtime are failed.
My question is:
1. Whether spring-cloud-dataflow-server-cloudfoundry can be used in SAP cloud foundry like I used?
2. When deploy stream in cloudfoundry using spring-cloud-dataflow-server-cloudfoundry dashboard, should I set any other necessary properties?
Thanks in advance.
Looking at the manifest.yml, it appears that org, space, and domain weren't replaced with SAP-CF specific values. Pay attention to the following note in the ref. guide.
Now we can configure the app. The following configuration is for Pivotal Web Services. You need to fill in {org}, {space}, {email} and {password} before running these commands.
If you have them replaced with your environment specific properties, the next step is to check the SCDF-server's logs. There will be particular details as to why the deployment failed if it did.
Now to answer your questions.
For #1, it is hard to say without logs or environment details. We don't actively test against SAP distribution of Cloud Foundry. As far as the distribution is compatible with Diego 1.7.1 and over, it should work. We also publish the CF compatible versions in project site. Perhaps this could be useful to compare the SAP CF environment and its foundation versions.
For #2, no, you don't need any other properties.
Related
I'm attempting to set up and run the basic code sample (python version) found here.
Unfortunately the provided login page fails to resolve (http://lms.valence.desire2learn.com/).
Is there a different sample host I can use to verify my setup?
You are correct; that test host has been retired and replaced by a new one that's associated with the Brightspace Developer Community. To use it, you will need to be a Developer Community member and use a test user provisioned on the new developer environment host.
I am using PCF and try to bulk /single import application using http url and found network is blocking extrnal http, is there option to upload my task jar without adding into http ?
Following url i am try to import
http://repo.spring.io/libs-snapshot/org/springframework/cloud/stream/app/spring-cloud-stream-app-descriptor/Celsius.BUILD-SNAPSHOT/spring-cloud-stream-app-descriptor-Celsius.BUILD-SNAPSHOT.stream-apps-kafka-10-maven
http://repo.spring.io/libs-release-local/org/springframework/cloud/stream/app/spring-cloud-stream-app-descriptor/Celsius.SR3/spring-cloud-stream-app-descriptor-Celsius.SR3.stream-apps-rabbit-maven
Yes, you can!
The HTTP URLs that we publish are nothing but a property file with key/value pairs of out-of-the-box application coordinates. You could download the file in your laptop, and use the 3rd choice from the page "Bulk import application coordinates from a property file.". Alternatively, from the same page, you could copy + paste the k/v pairs in the "Apps as Properties" text-area. These two options would allow the registration of application coordinates in SCDF's App registry.
However, at runtime, these applications will be resolved, downloaded, and deployed (by SCDF) as part of the stream/task deployments. That would mean, in a restricted environment, you may still run into the same connectivity problem.
For that reason, we have different other options in PCF to host/resolve application artifacts — see ref. docs. The SCDF App Tool is typically preferred by PCF customers.
Have deployed wso2is deployed on my k8s cluster using the Dockerfile mentioned in the https://github.com/wso2/docker-is/blob/5.7.x/dockerfiles/ubuntu/is-analytics/base/Dockerfile, its working fine.
Now the requirement has changed to have login stats (successful/unsuccessful/ failed attempts etc.) discover the Analytics support is the one use. But i am not quite sure how do i add this module into my Dockerfile?
Can someone list the various steps to install wso2is with analytics.
I have download the wso2is-analytics-5.7.0 zip, but i am not sure what else in the Dockerfile(from the link mentioned above) needs to change other than the :
"ARG WSO2_SERVER=wso2is-analytics"
Edited: so going once again the wso2is site : https://docs.wso2.com/display/IS570/Prerequisites+to+Publish+Statistics
Step 03: Configure Event Publishers , is these all optional? if already have wso2is deployed before?
because it says this "In a fresh WSO2 IS pack, you can view all the event publishers related to WSO2 IS Analytics in the /repository/deployment/server/eventpublishers directory."
Expected Result:
Have a working wso2is with analytics dasboard to track the login success/failure attempts.
thanks for your support, appreciate it!
Maurya (novice on wso2is)
The latest IS analytics-5.7.0, has different profiles, here, you need the following profiles,
Worker - Consume events from IS and process (This is consumed from event publishers in IS side, in the documentation, Step 3 and 4 are optional, that is documented for further understanding and not needed for deployment initially )
[1] https://hub.docker.com/r/wso2/wso2is-analytics-worker
Dashboard - This is to view the statistics
[2] https://hub.docker.com/r/wso2/wso2is-analytics-dashboard
[3] https://docs.wso2.com/display/IS570/Accessing+the+Analytics+Dashboard
I am absolutly new in WSO2 EI and I have the following doubt.
I am following this official tutorial to study the creation of a business process BPMN process):
https://docs.wso2.com/display/EI600/Quick+Start+Guide
If you go in this tutorial the second tab is refered to the Business Process Management tutorial, then I continue following the business process BPMN process street.
I followed all the tutorial and I think that I have understand the key concept of this example but I have some doubts about how to run this example.
At the end of the tutorial it does:
To create the Business Archive File (.bar), do the following.
On the Explorer, found in the upper left of your screen, navigate to Package Explorer.
Right click on the package and select “Create deployment artifacts”.
Navigate to the location of the source. You will find the .bar file inside the deployment folder.
So now I have my /HelloWorld/deployment/OrderApproval.bar file created.
But now what have I to do with this file?
I also have download the EI Server from here: /HelloWorld/deployment/OrderApproval.bar
Then I have configured it in my Eclipse Developer Studio EI setting it as a WSO2 server.
So my doubts are:
What exactly represent this OrderApproval.bar file? Is it a file representing a WSO2 EI executable project or what?
From what I have understand the previous example represents a work flow representing a BPMN process that implements a flow with some rules without that I need to write Java code (correct me if I am doing wrong assertion). But is it something executable or something that have to be integrated into some other application?
If it is an executable project what have I to do to run it and interact with it?
BPMN is process definition that can be run on BPS server. BPS server is part of WSO2 EI solution. You have to run it from [EI_HOME]/wso2/business-process/bin
On Linux/Mac OS: sh wso2server.sh
On Windows: wso2server.bat --run
Then, when it works correctly, open https : // [serverIP]:[port]/carbon/ and log into the EI-Business Process management console.
Next navigate to Home -> Manage -> Processes ->Add -> BPMN and upload your OrderApproval.bar file.
After uploading file, log into the BPMN-explorer at https : // [serverIP]:[port]/bpmn-explorer. On the "proecesses" tab you should see your process which can be run by action->start.
Try studie this example: https://docs.wso2.com/display/EI600/Creating+a+BPMN+Process
I am developing a service layer app which provides a catalog of webservices, then I am orchestrating them using OpenESB.
I create my BPELs importing external WSDL definitions using http://localhost:8080/services/myService?wsdl.
The problem is -- these BPELs strongly depend on this specific URL, and when I deploy on production server, my ESB layer stops working.
How can I make my BPELs independent of the specific endpoint? Can I refer the URIs to an external config file?
To do it you must create application configuration and application variable and add them on your http address. Example: "http://${MyHtttpAddress}:${MyHttpPort}/service1/myService?wsdl"/>.
Applications and variable are set up in the administrative console and can be changed for each environment.
Regards
Paul