Please help me, i'm new on cmdbuild and try to investigate this tools. But i stuck on the running of Shark.
I got the next errors in log :
shark.log:
2015-12-23 10:35:01,875: SharkEngineManager -> Working with 'org.enhydra.shark.CallbackUtil' implementation of core CallbackUtilities API
2015-12-23 10:35:01,875: SharkEngineManager -> Working with 'org.enhydra.shark.SharkObjectFactoryExt' implementation of core SharkObjectFactory API
2015-12-23 10:35:01,875: SharkEngineManager -> Working with 'org.enhydra.shark.StandardToolActivityHandler' implementation of core ToolAgentManager API
2015-12-23 10:35:01,875: SharkEngineManager -> Working with 'org.enhydra.shark.logging.StandardLoggingManager' implementation of Logging API
2015-12-23 10:35:02,441: StandardAssignmentManager -> Working with 'org.enhydra.shark.usergroup.DODSUserGroupManagerAdmin' implementation of UserGroup API
2015-12-23 10:35:02,517: StandardAssignmentManager -> Working with 'org.enhydra.shark.partmappersistence.DODSParticipantMappingAdmin' implementation of ParticipantMapping API
2015-12-23 10:35:02,517: SharkEngineManager -> Working with 'org.enhydra.shark.assignment.StandardAssignmentManager' implementation of Assignment API
2015-12-23 10:35:02,522: Process Cache configured - max. size is 100
2015-12-23 10:35:02,522: Resource Cache configured - max. size is 100
2015-12-23 10:35:02,522: SharkEngineManager -> Working with 'org.enhydra.shark.caching.LRUCacheMgr' implementation of Caching API
2015-12-23 10:35:02,676: <SqlQueryExecute><SqlQueryException>(<Replicated>:ERROR: relation "shkprocessstates" does not exist
Position: 32)
2015-12-23 10:35:02,684: <SqlQueryExecute><SqlQueryException>(<Replicated>:ERROR: current transaction is aborted, commands ignored until end of transaction block)
2015-12-23 10:35:02,687: DODSPersistentManager -> persisting String variables into BLOB for sizes greater than 4000
2015-12-23 10:35:02,687: SharkEngineManager -> Working with 'org.enhydra.shark.instancepersistence.DODSSelectivePersistenceManager' implementation of InstancePersistence API
2015-12-23 10:35:02,702: DODSGlobalPersistenceManager -> persisting String data into BLOB for sizes greater than 4000
2015-12-23 10:35:02,702: SharkEngineManager -> Working with 'org.enhydra.shark.globalpersistence.DODSGlobalPersistenceManager' implementation of GlobalPersistence API
2015-12-23 10:35:04,656: SharkEngineManager -> Working with 'org.cmdbuild.workflow.CMDBuildEventAuditManager' implementation of EventAudit API
2015-12-23 10:35:04,727: SharkEngineManager -> Working with 'org.enhydra.shark.repositorypersistence.DODSRepositoryPersistenceManager' implementation of RepositoryPersistence API
2015-12-23 10:35:04,730: SharkEngineManager -> Working with 'org.enhydra.shark.scripting.StandardScriptingManager' implementation of Scripting API
2015-12-23 10:35:04,731: SharkEngineManager -> Working without Security API implementation - SecurityManager is not specified.
2015-12-23 10:35:04,851: StandardToolAgentManager -> Working with 'org.enhydra.shark.appmappersistence.DODSApplicationMappingAdmin' implementation of ApplicationMapPersistence API
2015-12-23 10:35:04,851: SharkEngineManager -> Working with 'org.enhydra.shark.toolagent.StandardToolAgentManager' implementation of ToolAgentManager API
2015-12-23 10:35:04,851: SharkEngineManager -> Working without wfEngineInteroperability API implementation - WfEngineInteroperability implementation is not specified.
2015-12-23 10:35:04,862: SharkUtilities -> synchronizing XPDL cache
2015-12-23 10:35:04,878: <SqlQueryExecute><SqlQueryException>(<Replicated>:ERROR: current transaction is aborted, commands ignored until end of transaction block)
2015-12-23 10:35:04,878: Problem while initializing caches !!!
There is my catalina.out :
Shark will be configured from file /var/www/apache-tomcat-7.0.67/webapps/cmdbuild-shark-server-2.3.4/conf/Shark.conf
environmentType= tomcat
t= org.objectweb.jotm.Current#41e7458a
pre configure . . . . . .
Together Workflow Server 4.4-1
SharkEngineManager -> Shark engine is being initialized ...
com.lutris.dods.builder.generator.query.DataObjectException: Unknown SQLException: <SqlQueryExecute><SqlQueryException>(<Replicated>:ERROR: current transaction is aborted, commands ignored until end of transaction block)
at org.enhydra.shark.instancepersistence.data.ProcessStateQuery._runQuery(Unknown Source)
at org.enhydra.shark.instancepersistence.data.ProcessStateQuery.runQuery(Unknown Source)
at org.enhydra.shark.instancepersistence.data.ProcessStateQuery.getNextDO(Unknown Source)
at org.enhydra.shark.instancepersistence.data.ProcessStateDO.__refreshCache(Unknown Source)
at org.enhydra.shark.instancepersistence.data.ProcessStateDO.refreshCache(Unknown Source)
at org.enhydra.shark.instancepersistence.DODSPersistentManager.refreshDOCaches(Unknown Source)
at org.enhydra.shark.instancepersistence.DODSPersistentManager.setLDBName(Unknown Source)
at org.enhydra.shark.instancepersistence.DODSPersistentManager.configure(Unknown Source)
at org.enhydra.shark.instancepersistence.DODSSelectivePersistenceManager.configure(Unknown Source)
at org.enhydra.shark.SharkEngineManager.init(Unknown Source)
at org.enhydra.shark.SharkEngineManager.init(Unknown Source)
at org.enhydra.shark.SharkEngineManager.configure(Unknown Source)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at org.enhydra.shark.Shark._conf(Unknown Source)
at org.enhydra.shark.Shark.configure(Unknown Source)
at org.enhydra.shark.servlet.SharkInitServlet.init(SharkInitServlet.java:121)
at javax.servlet.GenericServlet.init(GenericServlet.java:158)
at org.apache.catalina.core.StandardWrapper.initServlet(StandardWrapper.java:1284)
at org.apache.catalina.core.StandardWrapper.loadServlet(StandardWrapper.java:1197)
at org.apache.catalina.core.StandardWrapper.load(StandardWrapper.java:1087)
at org.apache.catalina.core.StandardContext.loadOnStartup(StandardContext.java:5327)
at org.apache.catalina.core.StandardContext.startInternal(StandardContext.java:5617)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:150)
at org.apache.catalina.core.ContainerBase.addChildInternal(ContainerBase.java:901)
at org.apache.catalina.core.ContainerBase.addChild(ContainerBase.java:877)
at org.apache.catalina.core.StandardHost.addChild(StandardHost.java:652)
at org.apache.catalina.startup.HostConfig.deployWAR(HostConfig.java:1095)
at org.apache.catalina.startup.HostConfig$DeployWar.run(HostConfig.java:1957)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.sql.SQLException: <SqlQueryExecute><SqlQueryException>(<Replicated>:ERROR: current transaction is aborted, commands ignored until end of transaction block)
at com.lutris.dods.builder.generator.query.QueryBuilder.executeQuery(Unknown Source)
at org.enhydra.shark.instancepersistence.data.ProcessStateQuery.executeQuery(Unknown Source)
at com.lutris.appserver.server.sql.standard.StandardDBQuery.query(Unknown Source)
... 36 more
*** Caused by:
java.sql.SQLException: <SqlQueryExecute><SqlQueryException>(<Replicated>:ERROR: current transaction is aborted, commands ignored until end of transaction block)
at com.lutris.dods.builder.generator.query.QueryBuilder.executeQuery(Unknown Source)
at org.enhydra.shark.instancepersistence.data.ProcessStateQuery.executeQuery(Unknown Source)
at com.lutris.appserver.server.sql.standard.StandardDBQuery.query(Unknown Source)
at org.enhydra.shark.instancepersistence.data.ProcessStateQuery._runQuery(Unknown Source)
at org.enhydra.shark.instancepersistence.data.ProcessStateQuery.runQuery(Unknown Source)
at org.enhydra.shark.instancepersistence.data.ProcessStateQuery.getNextDO(Unknown Source)
at org.enhydra.shark.instancepersistence.data.ProcessStateDO.__refreshCache(Unknown Source)
at org.enhydra.shark.instancepersistence.data.ProcessStateDO.refreshCache(Unknown Source)
at org.enhydra.shark.instancepersistence.DODSPersistentManager.refreshDOCaches(Unknown Source)
at org.enhydra.shark.instancepersistence.DODSPersistentManager.setLDBName(Unknown Source)
at org.enhydra.shark.instancepersistence.DODSPersistentManager.configure(Unknown Source)
at org.enhydra.shark.instancepersistence.DODSSelectivePersistenceManager.configure(Unknown Source)
at org.enhydra.shark.SharkEngineManager.init(Unknown Source)
at org.enhydra.shark.SharkEngineManager.init(Unknown Source)
at org.enhydra.shark.SharkEngineManager.configure(Unknown Source)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at org.enhydra.shark.Shark._conf(Unknown Source)
at org.enhydra.shark.Shark.configure(Unknown Source)
at org.enhydra.shark.servlet.SharkInitServlet.init(SharkInitServlet.java:121)
at javax.servlet.GenericServlet.init(GenericServlet.java:158)
at org.apache.catalina.core.StandardWrapper.initServlet(StandardWrapper.java:1284)
at org.apache.catalina.core.StandardWrapper.loadServlet(StandardWrapper.java:1197)
at org.apache.catalina.core.StandardWrapper.load(StandardWrapper.java:1087)
at org.apache.catalina.core.StandardContext.loadOnStartup(StandardContext.java:5327)
at org.apache.catalina.core.StandardContext.startInternal(StandardContext.java:5617)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:150)
at org.apache.catalina.core.ContainerBase.addChildInternal(ContainerBase.java:901)
at org.apache.catalina.core.ContainerBase.addChild(ContainerBase.java:877)
at org.apache.catalina.core.StandardHost.addChild(StandardHost.java:652)
at org.apache.catalina.startup.HostConfig.deployWAR(HostConfig.java:1095)
at org.apache.catalina.startup.HostConfig$DeployWar.run(HostConfig.java:1957)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
*** Caused by:
java.sql.SQLException: <SqlQueryExecute><SqlQueryException>(<Replicated>:ERROR: current transaction is aborted, commands ignored until end of transaction block)
at com.lutris.dods.builder.generator.query.QueryBuilder.executeQuery(Unknown Source)
at org.enhydra.shark.instancepersistence.data.ProcessStateQuery.executeQuery(Unknown Source)
at com.lutris.appserver.server.sql.standard.StandardDBQuery.query(Unknown Source)
at org.enhydra.shark.instancepersistence.data.ProcessStateQuery._runQuery(Unknown Source)
at org.enhydra.shark.instancepersistence.data.ProcessStateQuery.runQuery(Unknown Source)
at org.enhydra.shark.instancepersistence.data.ProcessStateQuery.getNextDO(Unknown Source)
at org.enhydra.shark.instancepersistence.data.ProcessStateDO.__refreshCache(Unknown Source)
at org.enhydra.shark.instancepersistence.data.ProcessStateDO.refreshCache(Unknown Source)
at org.enhydra.shark.instancepersistence.DODSPersistentManager.refreshDOCaches(Unknown Source)
at org.enhydra.shark.instancepersistence.DODSPersistentManager.setLDBName(Unknown Source)
at org.enhydra.shark.instancepersistence.DODSPersistentManager.configure(Unknown Source)
at org.enhydra.shark.instancepersistence.DODSSelectivePersistenceManager.configure(Unknown Source)
at org.enhydra.shark.SharkEngineManager.init(Unknown Source)
at org.enhydra.shark.SharkEngineManager.init(Unknown Source)
at org.enhydra.shark.SharkEngineManager.configure(Unknown Source)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at org.enhydra.shark.Shark._conf(Unknown Source)
at org.enhydra.shark.Shark.configure(Unknown Source)
at org.enhydra.shark.servlet.SharkInitServlet.init(SharkInitServlet.java:121)
at javax.servlet.GenericServlet.init(GenericServlet.java:158)
at org.apache.catalina.core.StandardWrapper.initServlet(StandardWrapper.java:1284)
at org.apache.catalina.core.StandardWrapper.loadServlet(StandardWrapper.java:1197)
at org.apache.catalina.core.StandardWrapper.load(StandardWrapper.java:1087)
at org.apache.catalina.core.StandardContext.loadOnStartup(StandardContext.java:5327)
at org.apache.catalina.core.StandardContext.startInternal(StandardContext.java:5617)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:150)
at org.apache.catalina.core.ContainerBase.addChildInternal(ContainerBase.java:901)
at org.apache.catalina.core.ContainerBase.addChild(ContainerBase.java:877)
at org.apache.catalina.core.StandardHost.addChild(StandardHost.java:652)
at org.apache.catalina.startup.HostConfig.deployWAR(HostConfig.java:1095)
at org.apache.catalina.startup.HostConfig$DeployWar.run(HostConfig.java:1957)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
com.lutris.dods.builder.generator.query.DataObjectException: Unknown SQLException: <SqlQueryExecute><SqlQueryException>(<Replicated>:ERROR: current transaction is aborted, commands ignored until end of transaction block)
at org.enhydra.shark.repositorypersistence.data.XPDLQuery._runQuery(Unknown Source)
at org.enhydra.shark.repositorypersistence.data.XPDLQuery.runQuery(Unknown Source)
at org.enhydra.shark.repositorypersistence.data.XPDLQuery.getDOArray(Unknown Source)
at org.enhydra.shark.repositorypersistence.DODSRepositoryPersistenceManager.getAllXPDLs(Unknown Source)
at org.enhydra.shark.repositorypersistence.DODSRepositoryPersistenceManager.getExistingXPDLIds(Unknown Source)
at org.enhydra.shark.SharkUtilities.synchronizeXPDLCache(Unknown Source)
at org.enhydra.shark.SharkUtilities.restorePackages(Unknown Source)
at org.enhydra.shark.SharkEngineManager.initCaches(Unknown Source)
at org.enhydra.shark.SharkEngineManager.init(Unknown Source)
at org.enhydra.shark.SharkEngineManager.configure(Unknown Source)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at org.enhydra.shark.Shark._conf(Unknown Source)
at org.enhydra.shark.Shark.configure(Unknown Source)
at org.enhydra.shark.servlet.SharkInitServlet.init(SharkInitServlet.java:121)
at javax.servlet.GenericServlet.init(GenericServlet.java:158)
at org.apache.catalina.core.StandardWrapper.initServlet(StandardWrapper.java:1284)
at org.apache.catalina.core.StandardWrapper.loadServlet(StandardWrapper.java:1197)
at org.apache.catalina.core.StandardWrapper.load(StandardWrapper.java:1087)
at org.apache.catalina.core.StandardContext.loadOnStartup(StandardContext.java:5327)
at org.apache.catalina.core.StandardContext.startInternal(StandardContext.java:5617)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:150)
at org.apache.catalina.core.ContainerBase.addChildInternal(ContainerBase.java:901)
at org.apache.catalina.core.ContainerBase.addChild(ContainerBase.java:877)
at org.apache.catalina.core.StandardHost.addChild(StandardHost.java:652)
at org.apache.catalina.startup.HostConfig.deployWAR(HostConfig.java:1095)
at org.apache.catalina.startup.HostConfig$DeployWar.run(HostConfig.java:1957)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.sql.SQLException: <SqlQueryExecute><SqlQueryException>(<Replicated>:ERROR: current transaction is aborted, commands ignored until end of transaction block)
at com.lutris.dods.builder.generator.query.QueryBuilder.executeQuery(Unknown Source)
at org.enhydra.shark.repositorypersistence.data.XPDLQuery.executeQuery(Unknown Source)
at com.lutris.appserver.server.sql.standard.StandardDBQuery.query(Unknown Source)
... 34 more
*** Caused by:
java.sql.SQLException: <SqlQueryExecute><SqlQueryException>(<Replicated>:ERROR: current transaction is aborted, commands ignored until end of transaction block)
at com.lutris.dods.builder.generator.query.QueryBuilder.executeQuery(Unknown Source)
at org.enhydra.shark.repositorypersistence.data.XPDLQuery.executeQuery(Unknown Source)
at com.lutris.appserver.server.sql.standard.StandardDBQuery.query(Unknown Source)
at org.enhydra.shark.repositorypersistence.data.XPDLQuery._runQuery(Unknown Source)
at org.enhydra.shark.repositorypersistence.data.XPDLQuery.runQuery(Unknown Source)
at org.enhydra.shark.repositorypersistence.data.XPDLQuery.getDOArray(Unknown Source)
at org.enhydra.shark.repositorypersistence.DODSRepositoryPersistenceManager.getAllXPDLs(Unknown Source)
at org.enhydra.shark.repositorypersistence.DODSRepositoryPersistenceManager.getExistingXPDLIds(Unknown Source)
at org.enhydra.shark.SharkUtilities.synchronizeXPDLCache(Unknown Source)
at org.enhydra.shark.SharkUtilities.restorePackages(Unknown Source)
at org.enhydra.shark.SharkEngineManager.initCaches(Unknown Source)
at org.enhydra.shark.SharkEngineManager.init(Unknown Source)
at org.enhydra.shark.SharkEngineManager.configure(Unknown Source)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at org.enhydra.shark.Shark._conf(Unknown Source)
at org.enhydra.shark.Shark.configure(Unknown Source)
at org.enhydra.shark.servlet.SharkInitServlet.init(SharkInitServlet.java:121)
at javax.servlet.GenericServlet.init(GenericServlet.java:158)
at org.apache.catalina.core.StandardWrapper.initServlet(StandardWrapper.java:1284)
at org.apache.catalina.core.StandardWrapper.loadServlet(StandardWrapper.java:1197)
at org.apache.catalina.core.StandardWrapper.load(StandardWrapper.java:1087)
at org.apache.catalina.core.StandardContext.loadOnStartup(StandardContext.java:5327)
at org.apache.catalina.core.StandardContext.startInternal(StandardContext.java:5617)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:150)
at org.apache.catalina.core.ContainerBase.addChildInternal(ContainerBase.java:901)
at org.apache.catalina.core.ContainerBase.addChild(ContainerBase.java:877)
at org.apache.catalina.core.StandardHost.addChild(StandardHost.java:652)
at org.apache.catalina.startup.HostConfig.deployWAR(HostConfig.java:1095)
at org.apache.catalina.startup.HostConfig$DeployWar.run(HostConfig.java:1957)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
org.enhydra.shark.api.internal.repositorypersistence.RepositoryException: Unknown SQLException: <SqlQueryExecute><SqlQueryException>(<Replicated>:ERROR: current transaction is aborted, commands ignored until end of transaction block)
at org.enhydra.shark.repositorypersistence.DODSRepositoryPersistenceManager.getExistingXPDLIds(Unknown Source)
at org.enhydra.shark.SharkUtilities.synchronizeXPDLCache(Unknown Source)
at org.enhydra.shark.SharkUtilities.restorePackages(Unknown Source)
at org.enhydra.shark.SharkEngineManager.initCaches(Unknown Source)
at org.enhydra.shark.SharkEngineManager.init(Unknown Source)
at org.enhydra.shark.SharkEngineManager.configure(Unknown Source)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at org.enhydra.shark.Shark._conf(Unknown Source)
at org.enhydra.shark.Shark.configure(Unknown Source)
at org.enhydra.shark.servlet.SharkInitServlet.init(SharkInitServlet.java:121)
at javax.servlet.GenericServlet.init(GenericServlet.java:158)
at org.apache.catalina.core.StandardWrapper.initServlet(StandardWrapper.java:1284)
at org.apache.catalina.core.StandardWrapper.loadServlet(StandardWrapper.java:1197)
at org.apache.catalina.core.StandardWrapper.load(StandardWrapper.java:1087)
at org.apache.catalina.core.StandardContext.loadOnStartup(StandardContext.java:5327)
at org.apache.catalina.core.StandardContext.startInternal(StandardContext.java:5617)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:150)
at org.apache.catalina.core.ContainerBase.addChildInternal(ContainerBase.java:901)
at org.apache.catalina.core.ContainerBase.addChild(ContainerBase.java:877)
at org.apache.catalina.core.StandardHost.addChild(StandardHost.java:652)
at org.apache.catalina.startup.HostConfig.deployWAR(HostConfig.java:1095)
at org.apache.catalina.startup.HostConfig$DeployWar.run(HostConfig.java:1957)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Caused by: com.lutris.dods.builder.generator.query.DataObjectException: Unknown SQLException: <SqlQueryExecute><SqlQueryException>(<Replicated>:ERROR: current transaction is aborted, commands ignored until end of transaction block)
at org.enhydra.shark.repositorypersistence.data.XPDLQuery._runQuery(Unknown Source)
at org.enhydra.shark.repositorypersistence.data.XPDLQuery.runQuery(Unknown Source)
at org.enhydra.shark.repositorypersistence.data.XPDLQuery.getDOArray(Unknown Source)
at org.enhydra.shark.repositorypersistence.DODSRepositoryPersistenceManager.getAllXPDLs(Unknown Source)
... 30 more
Caused by: java.sql.SQLException: <SqlQueryExecute><SqlQueryException>(<Replicated>:ERROR: current transaction is aborted, commands ignored until end of transaction block)
at com.lutris.dods.builder.generator.query.QueryBuilder.executeQuery(Unknown Source)
at org.enhydra.shark.repositorypersistence.data.XPDLQuery.executeQuery(Unknown Source)
at com.lutris.appserver.server.sql.standard.StandardDBQuery.query(Unknown Source)
... 34 more
Shark -> shark engine initialization is finished, it lasted 3 [s]
Shark -> Shark ready and waiting ...
pre getConnection . . . . . .
pre commit . . . . . .
++++++++++++++++++++++++++++++++++++++ init - end
And sure my configs :
Shark.conf:
org.cmdbuild.ws.url=http://localhost:8080/cmdbuild/
org.cmdbuild.ws.username=workflow
org.cmdbuild.ws.password=cmdbuild
auth.conf:
auth.methods=DBAuthenticator
serviceuser.prigileged=workflow
Please help to understand, what i do wrong? Or what more information need to give?
P.S. All users and grants are setted like needed.
You need to ensure that tomcat is started as root. CMDBuild itself does not need root privileges to run but shark does.
In TOMCAT_HOME/cmdbuild/webapps/shark/META-INF/context.xml you should set your actual db name.
Related
I have installed Delta Lake package(delta-spark) in Zeppelin env and added delta-core dependency to spark - io.delta:delta-core_2.12:1.0.0.
Versions are as follows, I suppose I'm using the right delta-core version:
spark: 3.1.1.
scala: 2.12.10
Now, the delta module can be imported successfully. However, once I try to write or read data in delta format, it throws errors. Anyone knows what might go wrong here or how to fix it? Thanks!
For the following code:
%spark.pyspark
from delta import *
builder = SparkSession.builder.appName("MyApp") \
.config("spark.sql.extensions", "io.delta.sql.DeltaSparkSessionExtension") \
.config("spark.sql.catalog.spark_catalog", "org.apache.spark.sql.delta.catalog.DeltaCatalog")
spark = configure_spark_with_delta_pip(builder).getOrCreate()
data = spark.range(0, 5)
data.write.format("delta").save("hdfs://my-hdfs-namenode-0.my-hdfs-namenode.hdfs-explore.svc.cluster.local/tmp/delta-table-1")
It shows error happens in the last line with "java.lang.ClassNotFoundException: org.apache.spark.sql.delta.files.DelayedCommitProtocol":
Py4JJavaError: An error occurred while calling o122.save.
: org.apache.spark.SparkException: Job aborted.
at org.apache.spark.sql.execution.datasources.FileFormatWriter$.write(FileFormatWriter.scala:231)
at org.apache.spark.sql.delta.files.TransactionalWrite.$anonfun$writeFiles$1(TransactionalWrite.scala:192)
at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:103)
at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:163)
at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:90)
at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:772)
at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64)
at org.apache.spark.sql.delta.files.TransactionalWrite.writeFiles(TransactionalWrite.scala:163)
at org.apache.spark.sql.delta.files.TransactionalWrite.writeFiles$(TransactionalWrite.scala:142)
at org.apache.spark.sql.delta.OptimisticTransaction.writeFiles(OptimisticTransaction.scala:84)
at org.apache.spark.sql.delta.files.TransactionalWrite.writeFiles(TransactionalWrite.scala:135)
at org.apache.spark.sql.delta.files.TransactionalWrite.writeFiles$(TransactionalWrite.scala:134)
at org.apache.spark.sql.delta.OptimisticTransaction.writeFiles(OptimisticTransaction.scala:84)
at org.apache.spark.sql.delta.commands.WriteIntoDelta.write(WriteIntoDelta.scala:107)
at org.apache.spark.sql.delta.commands.WriteIntoDelta.$anonfun$run$1(WriteIntoDelta.scala:66)
at org.apache.spark.sql.delta.commands.WriteIntoDelta.$anonfun$run$1$adapted(WriteIntoDelta.scala:65)
at org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:187)
at org.apache.spark.sql.delta.commands.WriteIntoDelta.run(WriteIntoDelta.scala:65)
at org.apache.spark.sql.delta.sources.DeltaDataSource.createRelation(DeltaDataSource.scala:154)
at org.apache.spark.sql.execution.datasources.SaveIntoDataSourceCommand.run(SaveIntoDataSourceCommand.scala:46)
at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:70)
at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:68)
at org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:90)
at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1(SparkPlan.scala:180)
at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1(SparkPlan.scala:218)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:215)
at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:176)
at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:132)
at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:131)
at org.apache.spark.sql.DataFrameWriter.$anonfun$runCommand$1(DataFrameWriter.scala:989)
at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:103)
at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:163)
at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:90)
at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:772)
at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64)
at org.apache.spark.sql.DataFrameWriter.runCommand(DataFrameWriter.scala:989)
at org.apache.spark.sql.DataFrameWriter.saveToV1Source(DataFrameWriter.scala:438)
at org.apache.spark.sql.DataFrameWriter.saveInternal(DataFrameWriter.scala:409)
at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:293)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
at py4j.Gateway.invoke(Gateway.java:282)
at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
at py4j.commands.CallCommand.execute(CallCommand.java:79)
at py4j.GatewayConnection.run(GatewayConnection.java:238)
at java.lang.Thread.run(Thread.java:748)
Caused by: org.apache.spark.SparkException: Job aborted due to stage failure: Task 1 in stage 2.0 failed 4 times, most recent failure: Lost task 1.3 in stage 2.0 (TID 21) (10.80.3.38 executor 2): java.lang.ClassNotFoundException: org.apache.spark.sql.delta.files.DelayedCommitProtocol
at org.apache.spark.repl.ExecutorClassLoader.findClass(ExecutorClassLoader.scala:124)
at java.base/java.lang.ClassLoader.loadClass(Unknown Source)
at java.base/java.lang.ClassLoader.loadClass(Unknown Source)
at java.base/java.lang.Class.forName0(Native Method)
at java.base/java.lang.Class.forName(Unknown Source)
at org.apache.spark.serializer.JavaDeserializationStream$$anon$1.resolveClass(JavaSerializer.scala:68)
at java.base/java.io.ObjectInputStream.readNonProxyDesc(Unknown Source)
at java.base/java.io.ObjectInputStream.readClassDesc(Unknown Source)
at java.base/java.io.ObjectInputStream.readOrdinaryObject(Unknown Source)
at java.base/java.io.ObjectInputStream.readObject0(Unknown Source)
at java.base/java.io.ObjectInputStream.readArray(Unknown Source)
at java.base/java.io.ObjectInputStream.readObject0(Unknown Source)
at java.base/java.io.ObjectInputStream.defaultReadFields(Unknown Source)
at java.base/java.io.ObjectInputStream.readSerialData(Unknown Source)
at java.base/java.io.ObjectInputStream.readOrdinaryObject(Unknown Source)
at java.base/java.io.ObjectInputStream.readObject0(Unknown Source)
at java.base/java.io.ObjectInputStream.defaultReadFields(Unknown Source)
at java.base/java.io.ObjectInputStream.readSerialData(Unknown Source)
at java.base/java.io.ObjectInputStream.readOrdinaryObject(Unknown Source)
at java.base/java.io.ObjectInputStream.readObject0(Unknown Source)
at java.base/java.io.ObjectInputStream.readObject(Unknown Source)
at java.base/java.io.ObjectInputStream.readObject(Unknown Source)
at org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:76)
at org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:115)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:83)
at org.apache.spark.scheduler.Task.run(Task.scala:131)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:497)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1439)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:500)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
at java.base/java.lang.Thread.run(Unknown Source)
Caused by: java.lang.ClassNotFoundException: org.apache.spark.sql.delta.files.DelayedCommitProtocol
at java.base/java.lang.ClassLoader.findClass(Unknown Source)
at org.apache.spark.util.ParentClassLoader.findClass(ParentClassLoader.java:35)
at java.base/java.lang.ClassLoader.loadClass(Unknown Source)
at org.apache.spark.util.ParentClassLoader.loadClass(ParentClassLoader.java:40)
at java.base/java.lang.ClassLoader.loadClass(Unknown Source)
at org.apache.spark.repl.ExecutorClassLoader.findClass(ExecutorClassLoader.scala:109)
... 31 more
I guess the main issue here is that your cluster doesn't see the jar file of delta lake. If you added dependency using zeppelin spark interpreter, it has an impact only on Apache Zeppelin. You should upload the jar file of delta lake to /usr/lib/spark/jars/ in your cluster. The jar file can be downloaded from here.
I am generating a graphml file via python based on network analysis. This is working fine. The problem is that I am unable to read it using yED or cytoscape.
Here is the error that I am observing in yED.
java.io.IOException
at B.A.A.B.W.ă(Unknown Source)
at B.A.A.B.W.ā(Unknown Source)
at B.A.A.J.A.ā(Unknown Source)
at B.A.A.J.A.ā(Unknown Source)
at B.A.A.J.A$A.Ă(Unknown Source)
at B.A.A.J.A.ā(Unknown Source)
at B.A.A.K.D.ā(Unknown Source)
at B.A.A.K.j.actionPerformed(Unknown Source)
at javax.swing.AbstractButton.fireActionPerformed(AbstractButton.java:2022)
at javax.swing.AbstractButton$Handler.actionPerformed(AbstractButton.java:2348)
at javax.swing.DefaultButtonModel.fireActionPerformed(DefaultButtonModel.java:402)
at javax.swing.DefaultButtonModel.setPressed(DefaultButtonModel.java:259)
at javax.swing.AbstractButton.doClick(AbstractButton.java:376)
at com.apple.laf.ScreenMenuItem.actionPerformed(ScreenMenuItem.java:125)
at java.awt.MenuItem.processActionEvent(MenuItem.java:669)
at java.awt.MenuItem.processEvent(MenuItem.java:628)
at java.awt.MenuComponent.dispatchEventImpl(MenuComponent.java:351)
at java.awt.MenuComponent.dispatchEvent(MenuComponent.java:339)
at java.awt.EventQueue.dispatchEventImpl(EventQueue.java:761)
at java.awt.EventQueue.access$500(EventQueue.java:97)
at java.awt.EventQueue$3.run(EventQueue.java:709)
at java.awt.EventQueue$3.run(EventQueue.java:703)
at java.security.AccessController.doPrivileged(Native Method)
at java.security.ProtectionDomain$JavaSecurityAccessImpl.doIntersectionPrivilege(ProtectionDomain.java:76)
at java.security.ProtectionDomain$JavaSecurityAccessImpl.doIntersectionPrivilege(ProtectionDomain.java:86)
at java.awt.EventQueue$4.run(EventQueue.java:731)
at java.awt.EventQueue$4.run(EventQueue.java:729)
at java.security.AccessController.doPrivileged(Native Method)
at java.security.ProtectionDomain$JavaSecurityAccessImpl.doIntersectionPrivilege(ProtectionDomain.java:76)
at java.awt.EventQueue.dispatchEvent(EventQueue.java:728)
at java.awt.EventDispatchThread.pumpOneEventForFilters(EventDispatchThread.java:201)
at java.awt.EventDispatchThread.pumpEventsForFilter(EventDispatchThread.java:116)
at java.awt.EventDispatchThread.pumpEventsForHierarchy(EventDispatchThread.java:105)
at java.awt.EventDispatchThread.pumpEvents(EventDispatchThread.java:101)
at java.awt.EventDispatchThread.pumpEvents(EventDispatchThread.java:93)
at java.awt.EventDispatchThread.run(EventDispatchThread.java:82)
Caused by: java.lang.NumberFormatException: For input string: "10000000000"
at java.lang.NumberFormatException.forInputString(NumberFormatException.java:65)
at java.lang.Integer.parseInt(Integer.java:583)
at java.lang.Integer.valueOf(Integer.java:766)
at y.H.B.A$B.Ă(Unknown Source)
at y.H.B.B.X.ā(Unknown Source)
at y.H.B.B.Y.ā(Unknown Source)
at y.H.B.B.Y.ą(Unknown Source)
at y.H.B.B.Y.ā(Unknown Source)
at y.H.B.B.Y.Ă(Unknown Source)
at y.H.B.B.Y.ā(Unknown Source)
at y.H.B.B.Y.Ć(Unknown Source)
at y.H.B.B.Y.ā(Unknown Source)
at y.H.B.B.Y.Ą(Unknown Source)
at y.H.B.B.Y.ā(Unknown Source)
at y.H.B.B._.ā(Unknown Source)
at y.H.B.B._.ā(Unknown Source)
at y.H.B.A$13.ā(Unknown Source)
at y.H.B.A.đ(Unknown Source)
at y.H.B.A.ā(Unknown Source)
at y.H.Q.ā(Unknown Source)
at B.A.A.B.G.A.F.ā(Unknown Source)
at B.A.A.B.G.A.D.ā(Unknown Source)
at y.H.G.ā(Unknown Source)
at y.B.A.M.Đ(Unknown Source)
at y.B._.č(Unknown Source)
at y.B._.ĺ(Unknown Source)
at y.B._.ă(Unknown Source)
... 36 more
*
and here is my graphml file
http://pastebin.com/r5xuPTXX
I am new to the graphml format. Any help is appreciated.
The thing is Java parseInt returns a 32bit signed int which has a maximal value of 2,147,483,647, your value (maybe an node Id) is 10,000,000,000.
The parser code should return a long (a 64 bit integer) instead.
Networkx is not at fault here. If you cannot change the parser, maybe try to cast big values as string type instead of int.
I'm trying to figure out this error that starts after I refactor name of the project.(I tried reinstall Eclipse, create new workspace etc.)
My code was working well before that. If code is needed I can edit it later. Thanks for all advices.
Exception in Application start method
java.lang.reflect.InvocationTargetException
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at com.sun.javafx.application.LauncherImpl.launchApplicationWithArgs(Unknown Source)
at com.sun.javafx.application.LauncherImpl.launchApplication(Unknown Source)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at sun.launcher.LauncherHelper$FXHelper.main(Unknown Source)
Caused by: java.lang.RuntimeException: Exception in Application start method
at com.sun.javafx.application.LauncherImpl.launchApplication1(Unknown Source)
at com.sun.javafx.application.LauncherImpl.lambda$launchApplication$155(Unknown Source)
at java.lang.Thread.run(Unknown Source)
Caused by: java.lang.IllegalArgumentException: Invalid URL: Invalid URL or resource not found
at javafx.scene.image.Image.validateUrl(Unknown Source)
at javafx.scene.image.Image.<init>(Unknown Source)
at time.Run.start(Run.java:34)
at com.sun.javafx.application.LauncherImpl.lambda$launchApplication1$162(Unknown Source)
at com.sun.javafx.application.PlatformImpl.lambda$runAndWait$175(Unknown Source)
at com.sun.javafx.application.PlatformImpl.lambda$null$173(Unknown Source)
at java.security.AccessController.doPrivileged(Native Method)
at com.sun.javafx.application.PlatformImpl.lambda$runLater$174(Unknown Source)
at com.sun.glass.ui.InvokeLaterDispatcher$Future.run(Unknown Source)
at com.sun.glass.ui.win.WinApplication._runLoop(Native Method)
at com.sun.glass.ui.win.WinApplication.lambda$null$148(Unknown Source)
... 1 more
Caused by: java.lang.IllegalArgumentException: Invalid URL or resource not found
... 12 more
Exception running application time.Run
It compiled fine a few days ago but something i did must have caused the issue but so much was done i am not too sure.
Error log here: http://pastebin.com/P1JHqDBD
If anyone needs anything else to solve the issue, let me know!
I tried deleting the unitcache folder in my HTML5 folder. didnt help, same isse. also tried cleaning.
Thanks in advance
EDIT:
Nov 27, 2013 10:31:12 PM java.util.prefs.WindowsPreferences <init>
WARNING: Could not open/create prefs root node Software\JavaSoft\Prefs at root 0x80000002. Windows RegCreateKeyEx(...) returned error code 5.
[ERROR] Failure in unit cache map load.
java.util.concurrent.ExecutionException: java.lang.RuntimeException: Unable to read from byte cache
at java.util.concurrent.FutureTask.report(Unknown Source)
at java.util.concurrent.FutureTask.get(Unknown Source)
at com.google.gwt.dev.javac.PersistentUnitCache.awaitUnitCacheMapLoad(PersistentUnitCache.java:466)
at com.google.gwt.dev.javac.PersistentUnitCache.find(PersistentUnitCache.java:391)
at com.google.gwt.dev.javac.CompilationStateBuilder.addArchive(CompilationStateBuilder.java:365)
at com.google.gwt.dev.ArchivePreloader.preloadArchives(ArchivePreloader.java:65)
at com.google.gwt.dev.DevModeBase$UiBrowserWidgetHostImpl.createModuleSpaceHost(DevModeBase.java:106)
at com.google.gwt.dev.shell.OophmSessionHandler.loadModule(OophmSessionHandler.java:197)
at com.google.gwt.dev.shell.BrowserChannelServer.processConnection(BrowserChannelServer.java:526)
at com.google.gwt.dev.shell.BrowserChannelServer.run(BrowserChannelServer.java:364)
at java.lang.Thread.run(Unknown Source)
Caused by: java.lang.RuntimeException: Unable to read from byte cache
at com.google.gwt.dev.util.DiskCache.transferFromStream(DiskCache.java:171)
at com.google.gwt.dev.util.DiskCacheToken.readObject(DiskCacheToken.java:87)
at sun.reflect.GeneratedMethodAccessor3.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at java.io.ObjectStreamClass.invokeReadObject(Unknown Source)
at java.io.ObjectInputStream.readSerialData(Unknown Source)
at java.io.ObjectInputStream.readOrdinaryObject(Unknown Source)
at java.io.ObjectInputStream.readObject0(Unknown Source)
at java.io.ObjectInputStream.defaultReadFields(Unknown Source)
at java.io.ObjectInputStream.defaultReadObject(Unknown Source)
at com.google.gwt.dev.javac.CachedCompilationUnit.readObject(CachedCompilationUnit.java:204)
at sun.reflect.GeneratedMethodAccessor9.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at java.io.ObjectStreamClass.invokeReadObject(Unknown Source)
at java.io.ObjectInputStream.readSerialData(Unknown Source)
at java.io.ObjectInputStream.readOrdinaryObject(Unknown Source)
at java.io.ObjectInputStream.readObject0(Unknown Source)
at java.io.ObjectInputStream.readObject(Unknown Source)
at com.google.gwt.dev.javac.PersistentUnitCache.loadUnitMap(PersistentUnitCache.java:517)
at com.google.gwt.dev.javac.PersistentUnitCache.access$800(PersistentUnitCache.java:96)
at com.google.gwt.dev.javac.PersistentUnitCache$4.run(PersistentUnitCache.java:222)
at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source)
at java.util.concurrent.FutureTask.run(Unknown Source)
at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
... 1 more
Caused by: java.io.StreamCorruptedException: unexpected EOF in middle of data block
at java.io.ObjectInputStream$BlockDataInputStream.refill(Unknown Source)
at java.io.ObjectInputStream$BlockDataInputStream.read(Unknown Source)
at java.io.ObjectInputStream.read(Unknown Source)
at java.io.InputStream.read(Unknown Source)
at com.google.gwt.dev.util.DiskCache.transferFromStream(DiskCache.java:159)
... 27 more
I dont know what i did, i tried removing the war folder and put it back and suddenly it worked ....no idea :/
Currently I encounter some problems when I try to run Spark with Cassandra on standalone mode.
Initially, I run successfully with parameter mater="local[4]" in SparkContext.
Then, I try to move to standalone mode. What I used are:
Ubuntu: 12.04
Cassandra: 1.2.11
Spark: 0.8.0
Scala: 2.9.3
JDK: Oracle 1.6.0_35
Kryo: 2.21
At first, I got "unread block" error. As suggestion in other topic I change to use Kryo serializer and add Twitter Chill. Then, I get the " Failed to register spark.kryo.registrator " in my console and the Exception as below:
13/10/28 12:12:36 INFO cluster.ClusterTaskSetManager: Lost TID 0 (task 0.0:0)
13/10/28 12:12:36 INFO cluster.ClusterTaskSetManager: Loss was due to java.io.EOFException
java.io.EOFException
at org.apache.spark.serializer.KryoDeserializationStream.readObject(KryoSerializer.scala:109)
at org.apache.spark.broadcast.HttpBroadcast$.read(HttpBroadcast.scala:150)
at org.apache.spark.broadcast.HttpBroadcast.readObject(HttpBroadcast.scala:56)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at java.io.ObjectStreamClass.invokeReadObject(Unknown Source)
at java.io.ObjectInputStream.readSerialData(Unknown Source)
at java.io.ObjectInputStream.readOrdinaryObject(Unknown Source)
at java.io.ObjectInputStream.readObject0(Unknown Source)
at java.io.ObjectInputStream.defaultReadFields(Unknown Source)
at java.io.ObjectInputStream.readSerialData(Unknown Source)
at java.io.ObjectInputStream.readOrdinaryObject(Unknown Source)
at java.io.ObjectInputStream.readObject0(Unknown Source)
at java.io.ObjectInputStream.defaultReadFields(Unknown Source)
at java.io.ObjectInputStream.readSerialData(Unknown Source)
at java.io.ObjectInputStream.readOrdinaryObject(Unknown Source)
at java.io.ObjectInputStream.readObject0(Unknown Source)
at java.io.ObjectInputStream.readObject(Unknown Source)
at scala.collection.immutable.$colon$colon.readObject(List.scala:435)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at java.io.ObjectStreamClass.invokeReadObject(Unknown Source)
at java.io.ObjectInputStream.readSerialData(Unknown Source)
at java.io.ObjectInputStream.readOrdinaryObject(Unknown Source)
at java.io.ObjectInputStream.readObject0(Unknown Source)
at java.io.ObjectInputStream.defaultReadFields(Unknown Source)
at java.io.ObjectInputStream.readSerialData(Unknown Source)
at java.io.ObjectInputStream.readOrdinaryObject(Unknown Source)
at java.io.ObjectInputStream.readObject0(Unknown Source)
at java.io.ObjectInputStream.readObject(Unknown Source)
at org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:39)
at org.apache.spark.scheduler.ResultTask$.deserializeInfo(ResultTask.scala:61)
at org.apache.spark.scheduler.ResultTask.readExternal(ResultTask.scala:129)
at java.io.ObjectInputStream.readExternalData(Unknown Source)
at java.io.ObjectInputStream.readOrdinaryObject(Unknown Source)
at java.io.ObjectInputStream.readObject0(Unknown Source)
at java.io.ObjectInputStream.readObject(Unknown Source)
at org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:39)
at org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:61)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:153)
at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(Unknown Source)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
at java.lang.Thread.run(Unknown Source)
Someone also encountered the EOFException in spark before, the answer is not register the registrator correctly. I register the Registrator following the Spark guide. Registrator as below:
class MyRegistrator extends KryoRegistrator {
override def registerClasses(kryo: Kryo) {
kryo.register(classOf[org.apache.spark.rdd.RDD[(Map[String, ByteBuffer], Map[String, ByteBuffer])]])
kryo.register(classOf[String], 1)
kryo.register(classOf[Map[String, ByteBuffer]], 2)
}
}
And I also set the property just as the guide does.
System.setProperty("spark.serializer", "org.apache.spark.serializer.KryoSerializer")
System.setProperty("spark.kryo.registrator", "main.scala.MyRegistrator")
Can anyone give me some hints where I did wrong?
Thanks.
Based on my experience, the reasons to get the "EOFException" and the "data unread block" are the same. They lack some libraries when running on the clusters. The most wired thing is that I have add the libraries with "sbt assembly" in spark and the libraries are actually existing in the jars folder. But the spark still cannot find and load them successfully. Then I add the libraries in the spark context, it works. That means I need to transport the libraries to each nodes by specifying in the code.