I tried to compile chaincode sample and write my own one (https://github.com/hyperledger/fabric/tree/master/examples/chaincode/java/SimpleSample)
However, I can't get a shim-client-1.0. Do you have plans to make java sdk public?
I hope that “shim-client-1.0.jar” will be added to maven repository soon, nevertheless you can try to compile it using source code in “github.com/hyperledger/fabric/core/chaincode/shim/java”.
gradle build
and "shim-client-1.0.jar" will be available in github.com/hyperledger/fabric/core/chaincode/shim/java/build/lib folder.
Related
While trying to run the Google cloud dataflow Wordcount example in eclipse referenced here https://cloud.google.com/dataflow/docs/quickstarts/quickstart-java-eclipse
I am getting the following error:
An internal error occurred during: "Update Hierarchy".
Tried to create a TypeHierarchyPipelineOptionsHierarchy for a Java Project 'my project name' where no PipelineOptions type exists
This issue was also faced by somebody else as per the following stackoverflow link:
Eclipse: An internal error occurred during: "Update Hierarchy"
I tried the solution above, the project compiles but it does not run even after Force Update of Snapshots/Releases as explained above.
Based on my research of the problem it looks like google-cloud-dataflow-java-sdk-all-2.0.0-beta1.jar does not have the PipelineRunner class, which is causing the error. The 1.9.0 version of the same jar had those classes.
I cannot use 1.9.0 version of the jar directly because it causes other compilation errors e.g. package change to 'org.apache.beam' instead of 'com.google.cloud.dataflow'
Indeed, there's an issue in the older versions of the Google Cloud Dataflow plugin for Eclipse -- older versions are not forward-compatible with Dataflow SDKs 2.x series, but the project generation may still automatically create a project using the newest 2.x SDK.
To solve the problem, please upgrade to the newest version of the plugin.
Projects generated with versions 1.1.2 of the Dataflow Plugin for Eclipse and earlier will generate projects using the most recent archetype. With the release of Dataflow 2.0.0-beta1, the generated project will use the Apache Beam SDK as an underlying dependency. Modifying the project version to 1.9.0 or earlier will not modify the generated code, which causes the compilation failures you're experiencing.
Version 1.1.3 of the plugin will ensure that any generated project remains within the Dataflow namespace. Future versions will also work forwards-compatibly with Apache Beam. Version 1.1.3 should be available - in Eclipse, go to Help -> Check for Updates (possibly running Perform Setup Tasks to clear the cached current version) should make the updated plugin available.
We have published version 1.1.3 of the Dataflow Plugin for Eclipse; the quickstart example should now function. The project you've generated will not work until a future release of the Dataflow Plugin for Eclipse. Updating the plugin and re-running the quickstart should succeed.
I have successfully set up continuous integration process using TeamCity for non-SalesForce projects, but need help with automating SalesFoce manage packages installation.
To migrate metadata from source to target org, I used ant migration tool, but how to do in case of managed packages?
Questions:
How will TeamCity know that the new package has been created? I may watch the changes in source code tool, but not in the cloud where the created package is stored.
How to get the latest version of the package from the cloud? I need to compare the versions of the packages: the installed one and the new one
How to install the package?
Please help.
I checked SF documentation and found that SF API provides ability to install, upgrade and uninstall Managed Packages. Also it provides ability to run some Apex code after these actions. I think, you need to move in this way. Here is related documentation.
I created a sample Cloud Dataflow Java project using dataflow Eclipse plugin, I wanted to create a pipeline for DatastoreIO with namespace support, but i couldn't find the withNamespace() in the DatastoreIO class.
UPDATE: DatastoreIO namespace support is available in Cloud Dataflow SDK for Java, version 1.2.0 released this week.
DatastoreIO namespace support is available if you build the development version of the SDK from GitHub, however this version has not yet been released on Maven.
If you would like to use the DatastoreIO namespace support before that release happens, you can install the development version of the SDK using following the Getting Started instructions. The Eclipse project should then automatically pick up the new version, and if it does not you can edit the project pom.xml to manually specify the desired 1.1.1-SNAPSHOT version with these changes.
See also: https://github.com/GoogleCloudPlatform/DataflowJavaSDK/issues/63 (You do not normally need to post both places :).
I am new to Plugin development and hence facing this issue.
I have 2 plugin projects and both the projects are part of a feature project. I want to set up an ANT build which could build the feature project and publish it to an URL or File system.
Here is the scenario:
Plugin Project A isDependentOn -> Plugin Project B.
For both these project, I wrote build files which builds them properly.
Now I created a Feature project and included both these plugins into it. However I am not able to correctly write the ANT build file or may be I am not sure of the complete process.
Any help would be greatly appreciated.
Thanks in advance. Cheers !!!
Its better use Maven with Tycho to build eclipse features and plugins in headless mode.
Refer the link Eclipse Tycho and Using Tycho
I have an Android Project which uses droidText Library.
when I run the project using Eclipse Plugin, build well done.
However when I run it using ANT, the build faild and I got this error:
Pre-Dexing /droidText.0.4.jar -> droidText.0.4-39f8a99da14902e4ed96cdcc8f277d6e.jar
[DEBUG] Skipping watched dependency update for build
I'm prety sure that the problem is related to Android dependency and that's du to uses of duplicated classes in JAR file.
But I don't know how to fix that.
can somebody help me please?
The bug was related to the ANT version, before I had ant 1.7 but after I updated the it with ant 1.8.4. things went ok.