Stack: java.lang.IllegalArgumentException: No operation deployProjectJsonMergeOption found on MBean io.fabric8:type=ProjectDeployer - apache-camel

https://issues.jboss.org/browse/ENTESB-8039?focusedCommentId=13618981&page=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel#comment-13618981
In above ticket it mentioned as resolved but when i tried in jboss fuse 6.3 latest version, it is not updating the profile. As a result it thrown below error.
Stack: java.lang.IllegalArgumentException: No operation deployProjectJsonMergeOption found on MBean io.fabric8:type=ProjectDeployer

Looks like you are using a more recent version of the maven plugin than Fuse, hence the exception. The deployProjectJsonMergeOption method was introduced in 6.3 Roll Up 1. I'd suggest using the same version plugin/fuse or upgrading your Fuse installation.
HTH,
John.

Related

apache-camel 3.4.0 - Error executing command: No matching features for camel-catalog/0

Expected behaviour with camel version 2.15.2/2.24.1
Below url has feature as "camel-catalog" 
https://repo1.maven.org/maven2/org/apache/camel/karaf/apache-camel/2.15.2/apache-camel-2.15.2-features.xml
<feature name="camel-catalog" version="2.15.2" resolver="(obr)" start-level="50">
<feature version="2.15.2">camel-core
mvn:org.apache.camel.karaf/camel-karaf-commands-catalog/2.15.2
Exception while deploying feature ="camel-catalog" on Karaf Container with camel 3.x-
karaf#root()> feature:install camel-catalog
Error executing command: No matching features for camel-catalog/0
Query -
Upgraded camel version 3.4.0 now feature ="camel-catalog" is not there and but it require to install in Karaf container So what we can do to resolve this.
https://repo1.maven.org/maven2/org/apache/camel/karaf/apache-camel/3.4.0/apache-camel-3.4.0-features.xml
Any help/suggestion is most welcome.
The camel-catalog feature is not available as a feature in Camel 3, you dont need to install it, you get catalog installed via installing camel - features:install camel.
See the migration 2.x to 3.x guides: https://camel.apache.org/manual/latest/camel-3x-upgrade-guide.html

Payara 4.1.2.181 compatibility with Java 7

I can't run payara 4.1.2.181 with java 7 (Orracle jdk 1.7.0_80).
I have no problem with java 8 and the previous build of Payara (4.1.2.172) runs fine with Java 7. I couldn't find any mention about stopping support for Java 7 in this version.
The error I got is this one (server.log):
Caused by: org.osgi.framework.BundleException: Unresolved constraint in bundle fish.payara.microprofile.fault-tolerance.microprofile-fault-tolerance [149]: Unable to resolve 149.0: missing requirement [149.0] osgi.wiring.package; (&(osgi.wiring.package=org.eclipse.microprofile.config)(version>=1.0.0)(!(version>=2.0.0))) [caused by: Unable to resolve 223.0: missing requirement [223.0] osgi.ee; (&(osgi.ee=JavaSE)(version=1.8))]
Co you have any idea how to solve it?
Regards
Piotr
Payara stopped supporting JDK7 for public releases in 4.1.2.173, as explained in this blog post. It is still available for paying customers until 2023. So either upgrade to JDK8 or get Payara support.

Karaf feature:install throws Unsupported 'Bundle-ManifestVersion' value: 1

I am new to Karaf and trying to learn how to handle it.
On the way I tried to add Camunda features.
Like described on https://github.com/camunda/camunda-bpm-platform-osgi/tree/master/camunda-bpm-karaf-feature
at first, I added the repo:
feature:repo-add mvn:org.camunda.bpm.extension.osgi/camunda-bpm-karaf-feature/4.1.0/xml/features
then I tried to install them;
feature:install camunda-bpm-karaf-feature-full
unfortunately I got this Exception
org.osgi.framework.BundleException: Unable to build resource for mvn:xmlpull/xmlpull/1.1.3.1: Unsupported 'Bundle-ManifestVersion' value: 1
at org.apache.felix.utils.resource.ResourceBuilder.build(ResourceBuilder.java:82)
at org.apache.felix.utils.resource.ResourceBuilder.build(ResourceBuilder.java:67)
at org.apache.karaf.features.internal.region.SubsystemResolver.prepare(SubsystemResolver.java:180)
at org.apache.karaf.features.internal.service.Deployer.deploy(Deployer.java:379)
at org.apache.karaf.features.internal.service.FeaturesServiceImpl.doProvision(FeaturesServiceImpl.java:1025)
at org.apache.karaf.features.internal.service.FeaturesServiceImpl.lambda$doProvisionInThread$13(FeaturesServiceImpl.java:964)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Caused by: org.osgi.framework.BundleException: Unsupported 'Bundle-ManifestVersion' value: 1
at org.apache.felix.utils.resource.ResourceBuilder.doBuild(ResourceBuilder.java:90)
at org.apache.felix.utils.resource.ResourceBuilder.build(ResourceBuilder.java:80)
... 9 more
Error executing command: Unable to build resource for mvn:xmlpull/xmlpull/1.1.3.1: Unsupported 'Bundle-ManifestVersion' value: 1
I am using Karaf version 4.2.1
Does somebody know what I am doing wrong?
One of the features depend on xmlpull 1.1.3.1 which has Manifest-Version: 1.0 in MANIFEST.MF thus making it an OSGi R3 bundle.
Apache Felix supports bundles conforming the OSGi Release 4 (or newer) only (Manifest-Version: 2.0) which is why it rejects xmlpull. See the Felix source for reference.
If you control the source consider wrapping xmlpull and installing the wrapped bundle from the features. You can also play around with the Karaf console; for example, install -s wrap:mvn:xmlpull/xmlpull/1.1.3.1.

Flink 1.4 throws errors

just trying to migrate from flink 1.3 into 1.4 and getting this exception on
linux machine:
(not reproducing at windows).
i've import this package also:
// https://mvnrepository.com/artifact/org.apache.flink/flink-shaded-hadoop2
compile group: 'org.apache.flink', name: 'flink-shaded-hadoop2', version: '1.4.0'
any help?
at flink console:
TriggerWindow(TumblingProcessingTimeWindows(10000), ReducingStateDescriptor{serializer=org.apache.flink.api.java.typeutils.runtime.TupleSerializer#cb6c5dba, reduceFunction=com.clicktale.reducers.MetricsReducer#4e406694}, ProcessingTimeTrigger(), WindowedStream.reduce(WindowedStream.java:241)) -> Sink: Unnamed (1/1)
java.util.ServiceConfigurationError: org.apache.hadoop.fs.FileSystem: Provider org.apache.hadoop.fs.LocalFileSystem not a subtype
at java.util.ServiceLoader.fail(ServiceLoader.java:239)
at java.util.ServiceLoader.access$300(ServiceLoader.java:185)
at java.util.ServiceLoader$LazyIterator.nextService(ServiceLoader.java:376)
at java.util.ServiceLoader$LazyIterator.next(ServiceLoader.java:404)
at java.util.ServiceLoader$1.next(ServiceLoader.java:480)
at org.apache.hadoop.fs.FileSystem.loadFileSystems(FileSystem.java:2364)
at org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2375)
at org.apache.flink.runtime.fs.hdfs.HadoopFsFactory.create(HadoopFsFactory.java:99)
at org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:401)
at org.apache.flink.streaming.connectors.fs.bucketing.BucketingSink.createHadoopFileSystem(BucketingSink.java:1154)
at org.apache.flink.streaming.connectors.fs.bucketing.BucketingSink.initFileSystem(BucketingSink.java:411)
at org.apache.flink.streaming.connectors.fs.bucketing.BucketingSink.initializeState(BucketingSink.java:355)
at org.apache.flink.streaming.util.functions.StreamingFunctionUtils.tryRestoreFunction(StreamingFunctionUtils.java:178)
at org.apache.flink.streaming.util.functions.StreamingFunctionUtils.restoreFunctionState(StreamingFunctionUtils.java:160)
at org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator.initializeState(AbstractUdfStreamOperator.java:96)
at org.apache.flink.streaming.api.operators.AbstractStreamOperator.initializeState(AbstractStreamOperator.java:259)
at org.apache.flink.streaming.runtime.tasks.StreamTask.initializeOperators(StreamTask.java:694)
at org.apache.flink.streaming.runtime.tasks.StreamTask.initializeState(StreamTask.java:682)
at org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:253)
at org.apache.flink.runtime.taskmanager.Task.run(Task.java:718)
at java.lang.Thread.run(Thread.java:748)
I faced a similar (not specifically this, but dependencies related) issues migrating from 1.3 to 1.4.
In my case, I had to re-generate a fresh POM file using maven archetype and then add the needed dependencies one by one.
See Java Quickstart or Scala Quickstart.
Reason being that there has been a major rework on dependency structure. See Release notes for more information.
Note that Flink 1.4 will load any Hadoop jars found via the "hadoop classpath" shell command, and these will be first on the classpath. So if you have an incompatible version of Hadoop installed that the "hadoop" command points at, you can run into this kind of problem.

Apache Camel Classloader for Websphere

I am using Apache Camel with websphere. We have had some classloader issues.
I understand that Camel provides a websphere classloader but I cannot find an example in how to use it.
tried putting this in the apllicationContext.xml file
<bean id="WebsphereResolver" class="org.apache.camel.impl.WebSpherePackageScanClassResolver" />
but I got the error
Caused by: java.lang.NoSuchMethodException: org.apache.camel.impl.WebSpherePackageScanClassResolver.<init>()
What is correct format?
It turns out the problem was we were using an old version of fasterxml as defined in our pom.xml. Once we updated the pom.xml to use a new version of fasterxml, the problem went away. The apachexml bean uses fasterxml if it is available, but it assumes you are using a recent version.

Resources