how to generate OAS from kong api-gateway? - swagger

I have kong in its declarative mode and I am looking for a way to generate Open API Specification automatically. any ideas?
I have a kong.yaml file and I need oas based on my kong.yaml or kong admin

Related

How to override Kong ldap Auth Plugin configuration?

I am using Kong.yml file with DB less mode for configuring Kong gateway
I have configured ldap-auth plugin for all the services.
We need to authenticate user based on attribute sAMAccountName, so i need to modify kong/plugins/ldap-auth/access.lua file to support attributes
Where can i find below file in docker to modify
kong/plugins/ldap-auth/access.lua
Referred below link
github.com/Kong/kong/pull/5149/commits/d5752f71ff2c7a38b642bebab55f26eb73e286f9

Does the docker version of opennms have Rest API?

I am building a frontend-like for OpenNMS and chose to install OpenNMS in a docker container. I need to use the Rest API to further my project and when I try to send a request to
http://localhost:8980/opennms/rest, using the python requests library the return code is 404.
Does OpenNMS for docker not have Rest API or do I need to install it on my core system instead of docker.
P.S. This is my first time trying to do use the Rest API of an application.
imgur link : https://imgur.com/5UDPjRF
The URL you are calling is just the base URL for our rest resources. For Nodes as an example, you can test this with curl -u admin http://localhost:8980/opennms/rest/nodes. The REST API endpoints are described in the documentation here

Spring Boot: how to idiomatically configure Schema Registry Serdes in spring-kafka

Are there examples of configuring SpecificAvroSerdes (or any schema registry-based serdes - JsonSchema and Protobuf) in spring-kafka that allow leveraging some of the autoconfiguration (based on yaml or properties files).
There are a few similar questions in SO like How to use Spring-Kafka to read AVRO message with Confluent Schema registry?
But I want to be a bit specific on Kafka Streams serdes and declarative configuration of the Serdes.
Thank you
Looks like Kafka Streams can be configured with the default Serde.
See StreamsConfig.DEFAULT_KEY_SERDE_CLASS_CONFIG and StreamsConfig.DEFAULT_VALUE_SERDE_CLASS_CONFIG.
According Spring Boot convention we can provide any arbitrary properties from YAML file: https://docs.spring.io/spring-boot/docs/2.4.0/reference/html/spring-boot-features.html#boot-features-kafka-extra-props
So, probably your SpecificAvroSerdes can be configured in the application.properties like this:
spring.kafka.streams.properties[default.value.serde]=io.confluent.kafka.streams.serdes.avro.SpecificAvroSerdes

Jhipster can not select API first development using swagger-codegen

I'm going thru the Jhipster and can't select some option. How could I have it selectable? Why I can't select the 'API first development using swagger-codegen' option?
These are the console steps after running jhipster command:
Check your Internet connection.
If you are using an HTTP proxy, try this command: git config --global url."https://".insteadOf git://
? Which type of application would you like to create? Microservice application
? What is the base name of your application? service1
? As you are running in a microservice architecture, on which port would like your server to run? It should be unique to avoid port conflicts. 8086
? What is your default Java package name? com.comp.abc.service1
? Which service discovery server do you want to use? JHipster Registry (uses Eureka, provides Spring Cloud Config support and monitoring dashboards)
? Which type of authentication would you like to use? JWT authentication (stateless, with a token)
? Which type of database would you like to use? No database
? Would you like to use Maven or Gradle for building the backend? Maven
? Which other technologies would you like to use?
❯◯ API first development using swagger-codegen
◯ Asynchronous messages using Apache Kafka
To select answers on multiple choices questions you need to press the spacebar.

Dataflow 1.2.0 YAML configuration changes

Yesterday I upgraded my development environment to Spring Cloud Dataflow 1.2.0 and all of my sink/source apps dependencies.
I have two main issues:
javaOpts: -Xmx128m is not longer being picked up, so locally deployed apps have the default Xmx value.
Here is the format of my previously working Dataflow yaml config.
See full here: https://pastebin.com/p1JmLnLJ
spring:
cloud:
dataflow:
applicationProperties:
stream:
spring:
cloud:
deployer:
local:
javaOpts: -Xmx128m
Kafka config options like ssl.truststore.location etc. are not being read correctly. Another stackoverflow post indicated these must be marked like this "[ssl.truststore.location]". Is there some documented working yaml config or list of breaking changes with 1.2.0? The file based authentication block was also moved, and I was able to figure that one out.
Yes, It looks like a bug in Spring Cloud Local Deployer to consider the common application properties passed via args. Created https://github.com/spring-cloud/spring-cloud-deployer-local/issues/48 to track this.

Resources