-
Bug
-
Resolution: Done
-
Major
-
None
-
False
-
False
-
Undefined
-
-
2021 Week 31-33 (from Aug 2)
https://github.com/kiegroup/kogito-runtimes/pull/1379
it looks like there is an issue with marshalling
Stacktrace
java.lang.AssertionError:
1 expectation failed.
Expected status code <201> but was <500>.
Standard Output
2021-06-17 13:34:34,562 INFO [org.kie.kog.cod.api.uti.AddonsConfigDiscovery] (build-41) Performed addonsConfig discovery, found: AddonsConfig{usePersistence=true, useTracing=false, useMonitoring=false, usePrometheusMonitoring=false, useCloudEvents=false, useExplainability=false, useProcessSVG=false}
2021-06-17 13:34:34,857 INFO [org.kie.kog.cod.cor.uti.ApplicationGeneratorDiscovery] (build-41) Generator discovery performed, found [openapispecs, processes, rules, decisions, predictions]
2021-06-17 13:34:36,521 INFO [org.kie.kog.cod.api.uti.AddonsConfigDiscovery] (build-6) Performed addonsConfig discovery, found: AddonsConfig{usePersistence=true, useTracing=false, useMonitoring=false, usePrometheusMonitoring=false, useCloudEvents=false, useExplainability=false, useProcessSVG=false}
2021-06-17 13:34:37,039 INFO [org.tes.doc.DockerClientProviderStrategy] (build-19) Loaded org.testcontainers.dockerclient.EnvironmentAndSystemPropertyClientProviderStrategy from ~/.testcontainers.properties, will try it first
2021-06-17 13:34:37,764 INFO [org.tes.doc.DockerClientProviderStrategy] (build-19) Found Docker environment with Environment variables, system properties and defaults. Resolved dockerHost=unix:///var/run/docker.sock
2021-06-17 13:34:37,766 INFO [org.tes.DockerClientFactory] (build-19) Docker host IP address is localhost
2021-06-17 13:34:37,817 INFO [org.tes.DockerClientFactory] (build-19) Connected to docker:
Server Version: 20.10.6
API Version: 1.41
Operating System: Red Hat Enterprise Linux Server 7.9 (Maipo)
Total Memory: 15884 MB
2021-06-17 13:34:37,821 INFO [org.tes.uti.ImageNameSubstitutor] (build-19) Found configured ImageNameSubstitutor: Kogito Image Name Substitutor
2021-06-17 13:34:37,822 INFO [org.tes.uti.ImageNameSubstitutor] (build-19) Image name substitution will be performed by: Chained substitutor of 'DefaultImageNameSubstitutor (composite of 'ConfigurationFileImageNameSubstitutor' and 'PrefixingImageNameSubstitutor')' and then 'Kogito Image Name Substitutor'
2021-06-17 13:34:37,859 INFO [org.tes.uti.RegistryAuthLocator] (build-19) Failure when attempting to lookup auth config. Please ignore if you don't have images in an authenticated registry. Details: (dockerImageName: testcontainers/ryuk:0.3.1, configFile: /home/jenkins/.docker/config.json. Falling back to docker-java default behaviour. Exception message: /home/jenkins/.docker/config.json (No such file or directory)
2021-06-17 13:34:38,454 INFO [org.tes.DockerClientFactory] (build-19) Ryuk started - will monitor and terminate Testcontainers containers on JVM exit
2021-06-17 13:34:38,456 INFO [org.tes.uti.ImageNameSubstitutor] (build-19) Using library/mongo:4.0.10 as a substitute image for mongo:4.0.10 (using image substitutor: Kogito Image Name Substitutor)
2021-06-17 13:34:38,484 INFO [🐳 .0.10]] (build-19) Creating container for image: library/mongo:4.0.10
2021-06-17 13:34:38,533 INFO [🐳 .0.10]] (build-19) Starting container with ID: 2d5d189005406daf1775319d0a2dde533d6de75694318e697a324371324da5b4
2021-06-17 13:34:38,940 INFO [🐳 .0.10]] (build-19) Container library/mongo:4.0.10 is starting: 2d5d189005406daf1775319d0a2dde533d6de75694318e697a324371324da5b4
2021-06-17 13:34:39,674 INFO [🐳 .0.10]] (build-19) Container library/mongo:4.0.10 started in PT1.218642S
2021-06-17T17:34:44.239+0000 I CONTROL [main] Automatically disabling TLS 1.0, to force-enable TLS 1.0 specify --sslDisabledProtocols 'none'
2021-06-17T17:34:44.244+0000 I CONTROL [initandlisten] MongoDB starting : pid=1 port=27017 dbpath=/data/db 64-bit host=13136f02833f
2021-06-17T17:34:44.244+0000 I CONTROL [initandlisten] db version v4.0.10
2021-06-17T17:34:44.245+0000 I CONTROL [initandlisten] git version: c389e7f69f637f7a1ac3cc9fae843b635f20b766
2021-06-17T17:34:44.245+0000 I CONTROL [initandlisten] OpenSSL version: OpenSSL 1.0.2g 1 Mar 2016
2021-06-17T17:34:44.245+0000 I CONTROL [initandlisten] allocator: tcmalloc
2021-06-17T17:34:44.245+0000 I CONTROL [initandlisten] modules: none
2021-06-17T17:34:44.245+0000 I CONTROL [initandlisten] build environment:
2021-06-17T17:34:44.245+0000 I CONTROL [initandlisten] distmod: ubuntu1604
2021-06-17T17:34:44.245+0000 I CONTROL [initandlisten] distarch: x86_64
2021-06-17T17:34:44.245+0000 I CONTROL [initandlisten] target_arch: x86_64
2021-06-17T17:34:44.245+0000 I CONTROL [initandlisten] options: { net: { bindIpAll: true }, replication: { replSet: "docker-rs" } }
2021-06-17T17:34:44.245+0000 I STORAGE [initandlisten] wiredtiger_open config: create,cache_size=7430M,session_max=20000,eviction=(threads_min=4,threads_max=4),config_base=false,statistics=(fast),log=(enabled=true,archive=true,path=journal,compressor=snappy),file_manager=(close_idle_time=100000),statistics_log=(wait=0),verbose=(recovery_progress),
2021-06-17T17:34:44.856+0000 I STORAGE [initandlisten] WiredTiger message [1623951284:856741][1:0x7f8898406a80], txn-recover: Set global recovery timestamp: 0
2021-06-17T17:34:44.861+0000 I RECOVERY [initandlisten] WiredTiger recoveryTimestamp. Ts: Timestamp(0, 0)
2021-06-17T17:34:44.869+0000 I CONTROL [initandlisten]
2021-06-17T17:34:44.869+0000 I CONTROL [initandlisten] ** WARNING: Access control is not enabled for the database.
2021-06-17T17:34:44.869+0000 I CONTROL [initandlisten] ** Read and write access to data and configuration is unrestricted.
2021-06-17T17:34:44.869+0000 I CONTROL [initandlisten]
2021-06-17T17:34:44.870+0000 I CONTROL [initandlisten]
2021-06-17T17:34:44.870+0000 I CONTROL [initandlisten] ** WARNING: /sys/kernel/mm/transparent_hugepage/enabled is 'always'.
2021-06-17T17:34:44.870+0000 I CONTROL [initandlisten] ** We suggest setting it to 'never'
2021-06-17T17:34:44.870+0000 I CONTROL [initandlisten]
2021-06-17T17:34:44.870+0000 I CONTROL [initandlisten] ** WARNING: /sys/kernel/mm/transparent_hugepage/defrag is 'always'.
2021-06-17T17:34:44.870+0000 I CONTROL [initandlisten] ** We suggest setting it to 'never'
2021-06-17T17:34:44.870+0000 I CONTROL [initandlisten]
2021-06-17T17:34:44.873+0000 I STORAGE [initandlisten] createCollection: local.startup_log with generated UUID: e29208d4-5a5f-455b-8f05-97d404acde1d
2021-06-17T17:34:44.879+0000 I FTDC [initandlisten] Initializing full-time diagnostic data capture with directory '/data/db/diagnostic.data'
2021-06-17T17:34:44.880+0000 I STORAGE [initandlisten] createCollection: local.replset.oplogTruncateAfterPoint with generated UUID: 82a5c1fd-a9be-466a-80a5-16c0bc3f9014
2021-06-17T17:34:44.886+0000 I STORAGE [initandlisten] createCollection: local.replset.minvalid with generated UUID: 15d7bb89-cc81-44c4-b38e-e81a4805fcdd
2021-06-17T17:34:44.891+0000 I REPL [initandlisten] Did not find local voted for document at startup.
2021-06-17T17:34:44.891+0000 I REPL [initandlisten] Did not find local Rollback ID document at startup. Creating one.
2021-06-17T17:34:44.891+0000 I STORAGE [initandlisten] createCollection: local.system.rollback.id with generated UUID: bf915fd8-5989-4e0e-8792-16a92ad9ca0f
2021-06-17T17:34:44.896+0000 I REPL [initandlisten] Initialized the rollback ID to 1
2021-06-17T17:34:44.896+0000 I REPL [initandlisten] Did not find local replica set configuration document at startup; NoMatchingDocument: Did not find replica set configuration document in local.system.replset
2021-06-17T17:34:44.897+0000 I CONTROL [LogicalSessionCacheRefresh] Sessions collection is not set up; waiting until next sessions refresh interval: Replication has not yet been configured
2021-06-17T17:34:44.897+0000 I NETWORK [initandlisten] waiting for connections on port 27017
2021-06-17T17:34:44.897+0000 I CONTROL [LogicalSessionCacheReap] Sessions collection is not set up; waiting until next sessions reap interval: config.system.sessions does not exist
2021-06-17T17:34:45.057+0000 I NETWORK [listener] connection accepted from 127.0.0.1:60066 #1 (1 connection now open)
2021-06-17T17:34:45.058+0000 I NETWORK [conn1] received client metadata from 127.0.0.1:60066 conn1: { application: { name: "MongoDB Shell" }, driver: { name: "MongoDB Internal Client", version: "4.0.10" }, os: { type: "Linux", name: "Ubuntu", architecture: "x86_64", version: "16.04" } }
2021-06-17T17:34:45.061+0000 I COMMAND [conn1] initiate : no configuration specified. Using a default configuration for the set
2021-06-17T17:34:45.061+0000 I COMMAND [conn1] created this configuration for initiation : { _id: "docker-rs", version: 1, members: [ { _id: 0, host: "13136f02833f:27017" } ] }
2021-06-17T17:34:45.061+0000 I REPL [conn1] replSetInitiate admin command received from client
2021-06-17T17:34:45.062+0000 I REPL [conn1] replSetInitiate config object with 1 members parses ok
2021-06-17T17:34:45.062+0000 I REPL [conn1] ******
2021-06-17T17:34:45.062+0000 I REPL [conn1] creating replication oplog of size: 1640MB...
2021-06-17T17:34:45.062+0000 I STORAGE [conn1] createCollection: local.oplog.rs with generated UUID: ffa8d04a-3e7a-4fa1-885f-71cdc1c3c356
2021-06-17T17:34:45.077+0000 I STORAGE [conn1] Starting OplogTruncaterThread local.oplog.rs
2021-06-17T17:34:45.077+0000 I STORAGE [conn1] The size storer reports that the oplog contains 0 records totaling to 0 bytes
2021-06-17T17:34:45.077+0000 I STORAGE [conn1] Scanning the oplog to determine where to place markers for truncation
2021-06-17T17:34:45.092+0000 I REPL [conn1] ******
2021-06-17T17:34:45.092+0000 I STORAGE [conn1] createCollection: local.system.replset with generated UUID: 2dce4d8f-6129-4ac5-87fa-4e162e29a036
2021-06-17T17:34:45.099+0000 I STORAGE [conn1] createCollection: admin.system.version with provided UUID: a265ee16-4b2b-46a6-a1a9-6422546c3d27
2021-06-17T17:34:45.105+0000 I COMMAND [conn1] setting featureCompatibilityVersion to 4.0
2021-06-17T17:34:45.106+0000 I NETWORK [conn1] Skip closing connection for connection # 1
2021-06-17T17:34:45.106+0000 I REPL [conn1] New replica set config in use: { _id: "docker-rs", version: 1, protocolVersion: 1, writeConcernMajorityJournalDefault: true, members: [ { _id: 0, host: "13136f02833f:27017", arbiterOnly: false, buildIndexes: true, hidden: false, priority: 1.0, tags: {}, slaveDelay: 0, votes: 1 } ], settings: { chainingAllowed: true, heartbeatIntervalMillis: 2000, heartbeatTimeoutSecs: 10, electionTimeoutMillis: 10000, catchUpTimeoutMillis: -1, catchUpTakeoverDelayMillis: 30000, getLastErrorModes: {}, getLastErrorDefaults: { w: 1, wtimeout: 0 }, replicaSetId: ObjectId('60cb87b5d621d448c2a2f6e5') } }
2021-06-17T17:34:45.106+0000 I REPL [conn1] This node is 13136f02833f:27017 in the config
2021-06-17T17:34:45.106+0000 I REPL [conn1] transition to STARTUP2 from STARTUP
2021-06-17T17:34:45.106+0000 I REPL [conn1] Starting replication storage threads
2021-06-17T17:34:45.107+0000 I REPL [conn1] transition to RECOVERING from STARTUP2
2021-06-17T17:34:45.107+0000 I REPL [conn1] Starting replication fetcher thread
2021-06-17T17:34:45.107+0000 I REPL [conn1] Starting replication applier thread
2021-06-17T17:34:45.107+0000 I REPL [conn1] Starting replication reporter thread
2021-06-17T17:34:45.107+0000 I REPL [rsSync-0] Starting oplog application
2021-06-17T17:34:45.107+0000 I REPL [rsSync-0] transition to SECONDARY from RECOVERING
2021-06-17T17:34:45.107+0000 I REPL [rsSync-0] conducting a dry run election to see if we could be elected. current term: 0
2021-06-17T17:34:45.107+0000 I REPL [replexec-0] dry election run succeeded, running for election in term 1
2021-06-17T17:34:45.108+0000 I STORAGE [replexec-1] createCollection: local.replset.election with generated UUID: a64b9f75-ae9d-4d45-a786-75da48f65b24
2021-06-17T17:34:45.111+0000 I NETWORK [conn1] end connection 127.0.0.1:60066 (0 connections now open)
2021-06-17T17:34:45.114+0000 I REPL [replexec-1] election succeeded, assuming primary role in term 1
2021-06-17T17:34:45.114+0000 I REPL [replexec-1] transition to PRIMARY from SECONDARY
2021-06-17T17:34:45.114+0000 I REPL [replexec-1] Resetting sync source to empty, which was :27017
2021-06-17T17:34:45.114+0000 I REPL [replexec-1] Entering primary catch-up mode.
2021-06-17T17:34:45.114+0000 I REPL [replexec-1] Exited primary catch-up mode.
2021-06-17T17:34:45.114+0000 I REPL [replexec-1] Stopping replication producer
2021-06-17T17:34:45.274+0000 I NETWORK [listener] connection accepted from 127.0.0.1:60068 #2 (1 connection now open)
2021-06-17T17:34:45.275+0000 I NETWORK [conn2] received client metadata from 127.0.0.1:60068 conn2: { application: { name: "MongoDB Shell" }, driver: { name: "MongoDB Internal Client", version: "4.0.10" }, os: { type: "Linux", name: "Ubuntu", architecture: "x86_64", version: "16.04" } }
2021-06-17T17:34:47.109+0000 I STORAGE [rsSync-0] createCollection: config.transactions with generated UUID: bd4ab382-09cb-4037-b090-0fa57a501860
2021-06-17T17:34:47.116+0000 I REPL [rsSync-0] transition to primary complete; database writes are now permitted
2021-06-17T17:34:47.116+0000 I STORAGE [monitoring keys for HMAC] createCollection: admin.system.keys with generated UUID: b7b5b116-0c7d-4197-89ad-1ee57dd86ad7
2021-06-17T17:34:47.122+0000 I STORAGE [monitoring keys for HMAC] Triggering the first stable checkpoint. Initial Data: Timestamp(1623951285, 1) PrevStable: Timestamp(0, 0) CurrStable: Timestamp(1623951287, 4)
2021-06-17T17:34:47.198+0000 I NETWORK [conn2] end connection 127.0.0.1:60068 (0 connections now open)
Executing [/home/jenkins/workspace/KIE/kogito/pullrequest/kogito-runtimes.native.runtimes/kogito-runtimes/integration-tests/integration-tests-quarkus-processes-persistence/integration-tests-quarkus-processes-mongodb/target/integration-tests-quarkus-processes-mongodb-2.0.0-SNAPSHOT-runner, -Dquarkus.http.port=0, -Dquarkus.http.ssl-port=8444, -Dtest.url=http://localhost:0, -Dquarkus.log.file.path=/home/jenkins/workspace/KIE/kogito/pullrequest/kogito-runtimes.native.runtimes/kogito-runtimes/integration-tests/integration-tests-quarkus-processes-persistence/integration-tests-quarkus-processes-mongodb/target/quarkus.log, -Dquarkus.log.file.enable=true, -Dquarkus.mongodb.connection-string=mongodb://localhost:49266/test]
__ ____ __ _____ ___ __ ____ ______
--/ __ \/ / / / _ | / _ \/ //_/ / / / __/
-/ /_/ / /_/ / __ |/ , _/ ,< / /_/ /\ \
--\___\_\____/_/ |_/_/|_/_/|_|\____/___/
2021-06-17 13:34:47,332 WARN [io.qua.config] (main) Unrecognized configuration key "quarkus.jib.base.jvm.image" was provided; it will be ignored; verify that the dependency extension for this configuration is set or that you did not make a typo
2021-06-17 13:34:47,333 WARN [io.qua.config] (main) Unrecognized configuration key "quarkus.container.image.insecure" was provided; it will be ignored; verify that the dependency extension for this configuration is set or that you did not make a typo
2021-06-17 13:34:47,345 INFO [org.mon.dri.cluster] (main) Cluster created with settings {hosts=[localhost:49266], mode=SINGLE, requiredClusterType=UNKNOWN, serverSelectionTimeout='30000 ms'}
2021-06-17T17:34:47.349+0000 I NETWORK [listener] connection accepted from 172.17.0.1:48384 #3 (1 connection now open)
2021-06-17T17:34:47.349+0000 I NETWORK [listener] connection accepted from 172.17.0.1:48382 #4 (2 connections now open)
2021-06-17T17:34:47.349+0000 I NETWORK [conn3] received client metadata from 172.17.0.1:48384 conn3: { driver: { name: "mongo-java-driver|sync", version: "4.2.3" }, os: { type: "Linux", name: "Linux", architecture: "amd64", version: "3.10.0-1160.25.1.el7.x86_64" }, platform: "Java/Oracle Corporation/unknown-version" }
2021-06-17T17:34:47.349+0000 I NETWORK [conn4] received client metadata from 172.17.0.1:48382 conn4: { driver: { name: "mongo-java-driver|sync", version: "4.2.3" }, os: { type: "Linux", name: "Linux", architecture: "amd64", version: "3.10.0-1160.25.1.el7.x86_64" }, platform: "Java/Oracle Corporation/unknown-version" }
2021-06-17 13:34:47,350 INFO [org.mon.dri.connection] (cluster-rtt-ClusterId{value='60cb87b7e2721f3d580dce43', description='null'}-localhost:49266) Opened connection [connectionId{localValue:2, serverValue:4}] to localhost:49266
2021-06-17 13:34:47,351 INFO [org.mon.dri.connection] (cluster-ClusterId{value='60cb87b7e2721f3d580dce43', description='null'}-localhost:49266) Opened connection [connectionId{localValue:1, serverValue:3}] to localhost:49266
2021-06-17 13:34:47,351 INFO [org.mon.dri.cluster] (cluster-ClusterId{value='60cb87b7e2721f3d580dce43', description='null'}-localhost:49266) Monitor thread successfully connected to server with description ServerDescription{address=localhost:49266, type=REPLICA_SET_PRIMARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=7, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1540671, setName='docker-rs', canonicalAddress=13136f02833f:27017, hosts=[13136f02833f:27017], passives=[], arbiters=[], primary='13136f02833f:27017', tagSet=TagSet{[]}, electionId=7fffffff0000000000000001, setVersion=1, topologyVersion=null, lastWriteDate=Thu Jun 17 13:34:47 EDT 2021, lastUpdateTimeNanos=4706248349487}
2021-06-17 13:34:47,363 INFO [io.quarkus] (main) integration-tests-quarkus-processes-mongodb 2.0.0-SNAPSHOT native (powered by Quarkus 2.0.0.CR3) started in 0.041s. Listening on: http://0.0.0.0:39926
2021-06-17 13:34:47,363 INFO [io.quarkus] (main) Profile native activated.
2021-06-17 13:34:47,364 INFO [io.quarkus] (main) Installed features: [cdi, kogito-decisions, kogito-predictions, kogito-processes, kogito-rules, mongodb-client, resteasy, resteasy-jackson, servlet, smallrye-context-propagation]
2021-06-17 13:34:49,558 INFO [org.kie.kog.HelloService] (executor-thread-0) HelloService.hello invoked with params: [Tiago]
2021-06-17 13:34:49,559 ERROR [io.und.req.io] (executor-thread-0) Exception handling request 1efe9cee-66d0-491b-8619-45e06ded24ec-1 to /hello: org.jboss.resteasy.spi.UnhandledException: org.kie.kogito.serialization.process.ProcessInstanceMarshallerException: Error while marshalling process instance
at org.jboss.resteasy.core.ExceptionHandler.handleApplicationException(ExceptionHandler.java:106)
at org.jboss.resteasy.core.ExceptionHandler.handleException(ExceptionHandler.java:372)
at org.jboss.resteasy.core.SynchronousDispatcher.writeException(SynchronousDispatcher.java:218)
at org.jboss.resteasy.core.SynchronousDispatcher.invoke(SynchronousDispatcher.java:519)
at org.jboss.resteasy.core.SynchronousDispatcher.lambda$invoke$4(SynchronousDispatcher.java:261)
at org.jboss.resteasy.core.SynchronousDispatcher.lambda$preprocess$0(SynchronousDispatcher.java:161)
at org.jboss.resteasy.core.interception.jaxrs.PreMatchContainerRequestContext.filter(PreMatchContainerRequestContext.java:364)
at org.jboss.resteasy.core.SynchronousDispatcher.preprocess(SynchronousDispatcher.java:164)
at org.jboss.resteasy.core.SynchronousDispatcher.invoke(SynchronousDispatcher.java:247)
at org.jboss.resteasy.plugins.server.servlet.ServletContainerDispatcher.service(ServletContainerDispatcher.java:249)
at io.quarkus.resteasy.runtime.ResteasyFilter.doFilter(ResteasyFilter.java:35)
at io.undertow.servlet.core.ManagedFilter.doFilter(ManagedFilter.java:61)
at io.undertow.servlet.handlers.FilterHandler$FilterChainImpl.doFilter(FilterHandler.java:131)
at io.undertow.servlet.handlers.FilterHandler.handleRequest(FilterHandler.java:84)
at io.undertow.servlet.handlers.security.ServletSecurityRoleHandler.handleRequest(ServletSecurityRoleHandler.java:63)
at io.undertow.servlet.handlers.ServletChain$1.handleRequest(ServletChain.java:68)
at io.undertow.servlet.handlers.ServletDispatchingHandler.handleRequest(ServletDispatchingHandler.java:36)
at io.undertow.servlet.handlers.RedirectDirHandler.handleRequest(RedirectDirHandler.java:67)
at io.undertow.servlet.handlers.security.SSLInformationAssociationHandler.handleRequest(SSLInformationAssociationHandler.java:133)
at io.undertow.servlet.handlers.security.ServletAuthenticationCallHandler.handleRequest(ServletAuthenticationCallHandler.java:57)
at io.undertow.server.handlers.PredicateHandler.handleRequest(PredicateHandler.java:43)
at io.undertow.security.handlers.AbstractConfidentialityHandler.handleRequest(AbstractConfidentialityHandler.java:46)
at io.undertow.servlet.handlers.security.ServletConfidentialityConstraintHandler.handleRequest(ServletConfidentialityConstraintHandler.java:65)
at io.undertow.security.handlers.AuthenticationMechanismsHandler.handleRequest(AuthenticationMechanismsHandler.java:60)
at io.undertow.servlet.handlers.security.CachedAuthenticatedSessionHandler.handleRequest(CachedAuthenticatedSessionHandler.java:77)
at io.undertow.security.handlers.NotificationReceiverHandler.handleRequest(NotificationReceiverHandler.java:50)
at io.undertow.security.handlers.AbstractSecurityContextAssociationHandler.handleRequest(AbstractSecurityContextAssociationHandler.java:43)
at io.undertow.server.handlers.PredicateHandler.handleRequest(PredicateHandler.java:43)
at io.undertow.server.handlers.PredicateHandler.handleRequest(PredicateHandler.java:43)
at io.undertow.servlet.handlers.ServletInitialHandler.handleFirstRequest(ServletInitialHandler.java:247)
at io.undertow.servlet.handlers.ServletInitialHandler.access$100(ServletInitialHandler.java:56)
at io.undertow.servlet.handlers.ServletInitialHandler$2.call(ServletInitialHandler.java:111)
at io.undertow.servlet.handlers.ServletInitialHandler$2.call(ServletInitialHandler.java:108)
at io.undertow.servlet.core.ServletRequestContextThreadSetupAction$1.call(ServletRequestContextThreadSetupAction.java:48)
at io.undertow.servlet.core.ContextClassLoaderSetupAction$1.call(ContextClassLoaderSetupAction.java:43)
at io.quarkus.undertow.runtime.UndertowDeploymentRecorder$9$1.call(UndertowDeploymentRecorder.java:587)
at io.undertow.servlet.handlers.ServletInitialHandler.dispatchRequest(ServletInitialHandler.java:227)
at io.undertow.servlet.handlers.ServletInitialHandler.handleRequest(ServletInitialHandler.java:152)
at io.quarkus.undertow.runtime.UndertowDeploymentRecorder$1.handleRequest(UndertowDeploymentRecorder.java:119)
at io.undertow.server.Connectors.executeRootHandler(Connectors.java:290)
at io.undertow.server.DefaultExchangeHandler.handle(DefaultExchangeHandler.java:18)
at io.quarkus.undertow.runtime.UndertowDeploymentRecorder$5$1.run(UndertowDeploymentRecorder.java:413)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
at java.util.concurrent.FutureTask.run(FutureTask.java:264)
at io.quarkus.vertx.core.runtime.VertxCoreRecorder$14.runWith(VertxCoreRecorder.java:481)
at org.jboss.threads.EnhancedQueueExecutor$Task.run(EnhancedQueueExecutor.java:2442)
at org.jboss.threads.EnhancedQueueExecutor$ThreadBody.run(EnhancedQueueExecutor.java:1476)
at org.jboss.threads.DelegatingRunnable.run(DelegatingRunnable.java:29)
at org.jboss.threads.ThreadLocalResettingRunnable.run(ThreadLocalResettingRunnable.java:29)
at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.lang.Thread.run(Thread.java:829)
at com.oracle.svm.core.thread.JavaThreads.threadStartRoutine(JavaThreads.java:553)
at com.oracle.svm.core.posix.thread.PosixJavaThreads.pthreadStartRoutine(PosixJavaThreads.java:192)
Caused by: org.kie.kogito.serialization.process.ProcessInstanceMarshallerException: Error while marshalling process instance
at org.kie.kogito.serialization.process.ProcessInstanceMarshallerService.marshallProcessInstance(ProcessInstanceMarshallerService.java:111)
at org.kie.kogito.mongodb.MongoDBProcessInstances.updateStorage(MongoDBProcessInstances.java:113)
at org.kie.kogito.mongodb.MongoDBProcessInstances.create(MongoDBProcessInstances.java:98)
at org.kie.kogito.process.impl.AbstractProcessInstance.lambda$start$0(AbstractProcessInstance.java:226)
at org.kie.kogito.services.uow.ProcessInstanceWorkUnit.perform(ProcessInstanceWorkUnit.java:47)
at org.kie.kogito.services.uow.CollectingUnitOfWork.end(CollectingUnitOfWork.java:62)
at org.kie.kogito.services.uow.ManagedUnitOfWork.end(ManagedUnitOfWork.java:51)
at org.kie.kogito.services.uow.UnitOfWorkExecutor.executeInUnitOfWork(UnitOfWorkExecutor.java:34)
at org.kie.kogito.process.impl.ProcessServiceImpl.createProcessInstance(ProcessServiceImpl.java:58)
at org.kie.kogito.HelloResource.createResource_hello(HelloResource.java:72)
at org.kie.kogito.HelloResource_ClientProxy.createResource_hello(HelloResource_ClientProxy.zig:171)
at java.lang.reflect.Method.invoke(Method.java:566)
at org.jboss.resteasy.core.MethodInjectorImpl.invoke(MethodInjectorImpl.java:170)
at org.jboss.resteasy.core.MethodInjectorImpl.invoke(MethodInjectorImpl.java:130)
at org.jboss.resteasy.core.ResourceMethodInvoker.internalInvokeOnTarget(ResourceMethodInvoker.java:646)
at org.jboss.resteasy.core.ResourceMethodInvoker.invokeOnTargetAfterFilter(ResourceMethodInvoker.java:510)
at org.jboss.resteasy.core.ResourceMethodInvoker.lambda$invokeOnTarget$2(ResourceMethodInvoker.java:460)
at org.jboss.resteasy.core.interception.jaxrs.PreMatchContainerRequestContext.filter(PreMatchContainerRequestContext.java:364)
at org.jboss.resteasy.core.ResourceMethodInvoker.invokeOnTarget(ResourceMethodInvoker.java:462)
at org.jboss.resteasy.core.ResourceMethodInvoker.invoke(ResourceMethodInvoker.java:420)
at org.jboss.resteasy.core.ResourceMethodInvoker.invoke(ResourceMethodInvoker.java:394)
at org.jboss.resteasy.core.ResourceMethodInvoker.invoke(ResourceMethodInvoker.java:69)
at org.jboss.resteasy.core.SynchronousDispatcher.invoke(SynchronousDispatcher.java:492)
... 49 more
Caused by: org.kie.kogito.serialization.process.ProcessInstanceMarshallerException: No marshaller found for class java.lang.String
at org.kie.kogito.serialization.process.impl.ProtobufAbstractMarshallerContext.findMarshaller(ProtobufAbstractMarshallerContext.java:66)
at org.kie.kogito.serialization.process.impl.ProtobufAbstractMarshallerContext.findObjectMarshallerStrategyFor(ProtobufAbstractMarshallerContext.java:48)
at org.kie.kogito.serialization.process.impl.ProtobufVariableWriter.buildVariables(ProtobufVariableWriter.java:51)
at org.kie.kogito.serialization.process.impl.ProtobufProcessInstanceWriter.buildWorkflowContext(ProtobufProcessInstanceWriter.java:195)
at org.kie.kogito.serialization.process.impl.ProtobufProcessInstanceWriter.writeProcessInstance(ProtobufProcessInstanceWriter.java:146)
at org.kie.kogito.serialization.process.impl.ProtobufProcessInstanceMarshaller.writeProcessInstance(ProtobufProcessInstanceMarshaller.java:39)
at org.kie.kogito.serialization.process.ProcessInstanceMarshallerService.marshallProcessInstance(ProcessInstanceMarshallerService.java:108)
... 71 more
Request method: POST
Request URI: http://localhost:39926/hello
Proxy: <none>
Request params: <none>
Query params: <none>
Form params: <none>
Path params: <none>
Headers: Accept=*/*
Content-Type=application/json
Cookies: <none>
Multiparts: <none>
Body:
{
"var1": "Tiago"
}
HTTP/1.1 500 Internal Server Error
Content-Type: text/html;charset=UTF-8
Content-Length: 3180
<html lang="en">
<head>
<title>Internal Server Error - org.kie.kogito.serialization.process.ProcessInstanceMarshallerException: Error while marshalling process instanceError id 1efe9cee-66d0-491b-8619-45e06ded24ec-1</title>
<meta charset="utf-8"/>
<style>
html, body {
margin: 0;
padding: 0;
font-family: 'Open Sans', Helvetica, Arial, sans-serif;
font-size: 100%;
font-weight: 100;
line-height: 1.4;
}
html {
overflow-y: scroll;
}
body {
background: #f9f9f9;
}
.container {
width: 80%;
margin: 0 auto;
}
.content {
padding: 1em 0 1em 0;
}
header, .component-name {
background-color: #ad1c1c;
}
ul {
line-height: 1.5rem;
margin: 0.25em 0 0.25em 0;
}
.exception-message {
background: #be2828;
}
h1, h2 {
margin: 0;
padding: 0;
}
h1 {
font-size: 2rem;
color: #fff;
line-height: 3.75rem;
font-weight: 700;
padding: 0.4rem 0rem 0.4rem 0rem;
}
h2 {
font-size: 1.2rem;
color: rgba(255, 255, 255, 0.85);
line-height: 2.5rem;
font-weight: 400;
padding: 0.4rem 0rem 0.4rem 0rem;
}
.intro { font-size: 1.2rem;
font-weight: 400;
margin: 0.25em 0 1em 0;
}
h3 {
font-size: 1.2rem;
line-height: 2.5rem;
font-weight: 400;
color: #555;
margin: 0.25em 0 0.25em 0;
}
.trace, .resources {
background: #fff;
padding: 15px;
margin: 15px auto;
border: 1px solid #ececec;
}
.trace {
overflow-y: scroll;
}
.hidden {
display: none;
}
pre {
white-space: pre;
font-family: Consolas, Monaco, Menlo, "Ubuntu Mono", "Liberation Mono", monospace;
font-size: 12px;
line-height: 1.5;
color: #555;
}
</style>
<script>
function toggleStackTraceOrder() {
var stElement = document.getElementById('stacktrace');
var current = stElement.getAttribute('data-current-setting');
if (current == 'original-stacktrace') {
var reverseOrder = document.getElementById('reversed-stacktrace');
stElement.innerHTML = reverseOrder.innerHTML;
stElement.setAttribute('data-current-setting', 'reversed-stacktrace');
} else {
var originalOrder = document.getElementById('original-stacktrace');
stElement.innerHTML = originalOrder.innerHTML;
stElement.setAttribute('data-current-setting', 'original-stacktrace');
}
return;
}
function showDefaultStackTraceOrder() {
var reverseOrder = document.getElementById('reversed-stacktrace');
var stElement = document.getElementById('stacktrace');
if (reverseOrder == null || stElement == null) {
return;
}
// default to reverse ordered stacktrace
stElement.innerHTML = reverseOrder.innerHTML;
stElement.setAttribute('data-current-setting', 'reversed-stacktrace');
return;
}
</script>
</head>
<body onload="showDefaultStackTraceOrder()">
<header>
<h1 class="container">Internal Server Error</h1>
<div class="exception-message">
<h2 class="container">org.kie.kogito.serialization.process.ProcessInstanceMarshallerException: Error while marshalling process instanceError id 1efe9cee-66d0-491b-8619-45e06ded24ec-1</h2>
</div>
</header>
<div class="container content"/>
</body>
</html>
- relates to
-
KOGITO-5360 Native Image tests are broken in PostgreSQL Persistence
-
- Resolved
-
-
KOGITO-5361 Native Image tests are not configured correctly on persistence addons
-
- Resolved
-