ceph cluster break down after power failure, how can I save my data? - ceph

I have a 4 node ceph cluster deployed with cephadm on rocky9.0, ceph version 17.2.5. podman version 4.1.1
The entire cluster was broke down after an unexpected power failure.
cluster conf
# minimal ceph.conf for bf74b922-5d17-11ed-977a-525400b60738
[global]
fsid = bf74b922-5d17-11ed-977a-525400b60738
mon_host = [v2:192.168.129.190:3300/0,v1:192.168.129.190:6789/0] [v2:192.168.129.198:3300/0,v1:192.168.129.198:6789/0] [v2:192.168.129.232:3300/0,v1:192.168.129.232:6789/0]
[mon.ceph01]
public network = 192.168.129.0/24
check cluster status
ceph -s time out error message:
2023-02-02T08:06:05.037+0000 7f663699d700 0 monclient(hunting): authenticate timed out after 300
check osd service
No output running podman logs ceph-bf74b922-5d17-11ed-977a-525400b60738-osd-2
dashboard available
also checked :
hard drive, all with good health.
selinux and firewalld, disabled
podman, startus ok
network, status ok
DNS, status ok
NTP, status ok
Most containers will be restarted about every 50 minutes.
I have no idea about how to bring it back online. Is there any kind of checklist recomanded to prevent further damage, what should I do with this, how can I fix it.
check mon log
mon.ceph01 error log is like this:
Feb 13 15:38:59 ceph01.pinkcloud.localdomain ceph-bf74b922-5d17-11ed-977a-525400b60738-mon-ceph01[2403]: -1> 2023-02-13T07:38:59.399+0000 7f49cfb95880 -1 /home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos8/DIST/centos8/MACHINE_SIZE/gigantic/release/17.2.5/rpm/el8/BUILD/ceph-17.2.5/src/kv/RocksDBStore.cc: In function 'virtual int RocksDBStore::get(const string&, const string&, ceph::bufferlist*)' thread 7f49cfb95880 time 2023-02-13T07:38:59.399608+0000
Feb 13 15:38:59 ceph01.pinkcloud.localdomain ceph-bf74b922-5d17-11ed-977a-525400b60738-mon-ceph01[2403]: /home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos8/DIST/centos8/MACHINE_SIZE/gigantic/release/17.2.5/rpm/el8/BUILD/ceph-17.2.5/src/kv/RocksDBStore.cc: 1899: ceph_abort_msg("block checksum mismatch: stored = 3293112312, computed = 2419631843 in /var/lib/ceph/mon/ceph-ceph01/store.db/214973.sst offset 18759843 size 3815")
Feb 13 15:38:59 ceph01.pinkcloud.localdomain ceph-bf74b922-5d17-11ed-977a-525400b60738-mon-ceph01[2403]:
Feb 13 15:38:59 ceph01.pinkcloud.localdomain ceph-bf74b922-5d17-11ed-977a-525400b60738-mon-ceph01[2403]: ceph version 17.2.5 (98318ae89f1a893a6ded3a640405cdbb33e08757) quincy (stable)
Feb 13 15:38:59 ceph01.pinkcloud.localdomain ceph-bf74b922-5d17-11ed-977a-525400b60738-mon-ceph01[2403]: 1: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&)+0xd7) [0x7f49cf062a9f]
Feb 13 15:38:59 ceph01.pinkcloud.localdomain ceph-bf74b922-5d17-11ed-977a-525400b60738-mon-ceph01[2403]: 2: (RocksDBStore::get(std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, ceph::buffer::v15_2_0::list*)+0x3b3) [0x55960fc70453]
Feb 13 15:38:59 ceph01.pinkcloud.localdomain ceph-bf74b922-5d17-11ed-977a-525400b60738-mon-ceph01[2403]: 3: main()
Feb 13 15:38:59 ceph01.pinkcloud.localdomain ceph-bf74b922-5d17-11ed-977a-525400b60738-mon-ceph01[2403]: 4: __libc_start_main()
Feb 13 15:38:59 ceph01.pinkcloud.localdomain ceph-bf74b922-5d17-11ed-977a-525400b60738-mon-ceph01[2403]: 5: _start()
Feb 13 15:38:59 ceph01.pinkcloud.localdomain ceph-bf74b922-5d17-11ed-977a-525400b60738-mon-ceph01[2403]:
Feb 13 15:38:59 ceph01.pinkcloud.localdomain ceph-bf74b922-5d17-11ed-977a-525400b60738-mon-ceph01[2403]: 0> 2023-02-13T07:38:59.401+0000 7f49cfb95880 -1 *** Caught signal (Aborted) **
Feb 13 15:38:59 ceph01.pinkcloud.localdomain ceph-bf74b922-5d17-11ed-977a-525400b60738-mon-ceph01[2403]: in thread 7f49cfb95880 thread_name:ceph-mon
Feb 13 15:38:59 ceph01.pinkcloud.localdomain ceph-bf74b922-5d17-11ed-977a-525400b60738-mon-ceph01[2403]:
Feb 13 15:38:59 ceph01.pinkcloud.localdomain ceph-bf74b922-5d17-11ed-977a-525400b60738-mon-ceph01[2403]: ceph version 17.2.5 (98318ae89f1a893a6ded3a640405cdbb33e08757) quincy (stable)
Feb 13 15:38:59 ceph01.pinkcloud.localdomain ceph-bf74b922-5d17-11ed-977a-525400b60738-mon-ceph01[2403]: 1: /lib64/libpthread.so.0(+0x12cf0) [0x7f49cd00dcf0]
Feb 13 15:38:59 ceph01.pinkcloud.localdomain ceph-bf74b922-5d17-11ed-977a-525400b60738-mon-ceph01[2403]: 2: gsignal()
Feb 13 15:38:59 ceph01.pinkcloud.localdomain ceph-bf74b922-5d17-11ed-977a-525400b60738-mon-ceph01[2403]: 3: abort()
Feb 13 15:38:59 ceph01.pinkcloud.localdomain ceph-bf74b922-5d17-11ed-977a-525400b60738-mon-ceph01[2403]: 4: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&)+0x197) [0x7f49cf062b5f]
Feb 13 15:38:59 ceph01.pinkcloud.localdomain ceph-bf74b922-5d17-11ed-977a-525400b60738-mon-ceph01[2403]: 5: (RocksDBStore::get(std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, ceph::buffer::v15_2_0::list*)+0x3b3) [0x55960fc70453]
Feb 13 15:38:59 ceph01.pinkcloud.localdomain ceph-bf74b922-5d17-11ed-977a-525400b60738-mon-ceph01[2403]: 6: main()
Feb 13 15:38:59 ceph01.pinkcloud.localdomain ceph-bf74b922-5d17-11ed-977a-525400b60738-mon-ceph01[2403]: 7: __libc_start_main()
Feb 13 15:38:59 ceph01.pinkcloud.localdomain ceph-bf74b922-5d17-11ed-977a-525400b60738-mon-ceph01[2403]: 8: _start()
Feb 13 15:38:59 ceph01.pinkcloud.localdomain ceph-bf74b922-5d17-11ed-977a-525400b60738-mon-ceph01[2403]: NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to interpret this.
Feb 13 15:38:59 ceph01.pinkcloud.localdomain ceph-bf74b922-5d17-11ed-977a-525400b60738-mon-ceph01[2403]:
Feb 13 15:38:59 ceph01.pinkcloud.localdomain podman[2437]: 2023-02-13 15:38:59.680818197 +0800 CST m=+0.070475611 container died da69765578771a381ad9d84870d97a167b601f75a2f13fde9fb55f958d1811bc (image=quay.io/ceph/ceph#sha256:0560b16bec6e84345f29fb6693cd2430884e6efff16a95d5bdd0bb06d7661c45, name=ceph-bf74b922-5d17-11ed-977a-525400b60738-mon-ceph01)
Feb 13 15:38:59 ceph01.pinkcloud.localdomain podman[2437]: 2023-02-13 15:38:59.717834149 +0800 CST m=+0.107491608 container remove da69765578771a381ad9d84870d97a167b601f75a2f13fde9fb55f958d1811bc (image=quay.io/ceph/ceph#sha256:0560b16bec6e84345f29fb6693cd2430884e6efff16a95d5bdd0bb06d7661c45, name=ceph-bf74b922-5d17-11ed-977a-525400b60738-mon-ceph01, io.openshift.tags=base centos centos-stream, release=943, maintainer=Guillaume Abrioux <gabrioux#redhat.com>, url=https://access.redhat.com/containers/#/registry.access.redhat.com/ubi8/images/8.6-943, vcs-type=git, distribution-scope=public, io.k8s.description=The Universal Base Image is designed and engineered to be the base layer for all of your containerized applications, middleware and utilities. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.expose-services=, ceph=True, io.k8s.display-name=CentOS Stream 8, io.buildah.version=1.19.8, build-date=2022-09-06T17:25:04.845275, GIT_COMMIT=f77ca5de7910f1e3de260a1218c757954afd8327, vendor=Red Hat, Inc., com.redhat.build-host=cpt-1002.osbs.prod.upshift.rdu2.redhat.com, GIT_CLEAN=True, RELEASE=HEAD, com.redhat.component=centos-stream-container, description=CentOS Stream is a continuously delivered distro that tracks just ahead of Red Hat Enterprise Linux development. This image takes the Red Hat UBI and layers on content from CentOS Stream, architecture=x86_64, name=centos-stream, version=8, GIT_BRANCH=HEAD, CEPH_POINT_RELEASE=-17.2.5, vcs-ref=f1ee6e37554363ec55e0035aba1a693d3627fdeb, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.license_terms=https://centos.org/legal/licensing-policy/, summary=Provides a CentOS Stream container based on the Red Hat Universal Base Image
tried restart ceph.target, but it doesn't help anything.
how to repair this "block checksum mismatch" thing?

If all servers went down at the same time the data should be fine except for some pending writes. But the overall integrity should be intact. You need to start the MONs first so they form a quorum and other services and clients can connect. If the MONs fail to start check the MON logs with cephadm logs --name mon.<YOUR_MON>, and the same for other services. Basically, try to start in the reverse order of a regular shutdown procedure, one example from the SUSE docs.

Related

CentOS 8 freeze on gcloud. How to debug?

I run a Google Compute Engine (GCE) instance with CentOS 8 on Google Cloud Platform. The problem is that this instance freezes at random times and the only way to make it work is to stop the instance and start it again. I tried to recreate an instance, this doesn't help.
Here are the screenshots from the GCE monitoring page (freeze from 21:04 Feb 9 to 09:29 Feb 10):
CPU utilization + Network bytes
RAM + Disk space utilization
Network packets
It's a 2vCPU instance with 2Gb RAM. It runs 2 docker containers (1 for Apache, 1 for FastAPI), it also runs 4 cron jobs every minute. When the machine freezes I can't ssh to it, I can't access any webpage, none of the 4 cron jobs sends any data to a logging server.
Any ideas how I can debug this issue?
Here is the /var/log/messages dump in case it's useful:
Feb 9 21:03:07 instance-prod-2 systemd[285938]: Stopped target Basic System.
Feb 9 21:03:07 instance-prod-2 systemd[285938]: Stopped target Paths.
Feb 9 21:03:07 instance-prod-2 systemd[285938]: Stopped target Timers.
Feb 9 21:03:07 instance-prod-2 systemd[285938]: grub-boot-success.timer: Succeeded.
Feb 9 21:03:07 instance-prod-2 systemd[285938]: Stopped Mark boot as successful after the user session has run 2 minutes.
Feb 9 21:03:07 instance-prod-2 systemd[285938]: Stopped target Sockets.
Feb 9 21:03:07 instance-prod-2 systemd[285938]: dbus.socket: Succeeded.
Feb 9 21:03:07 instance-prod-2 systemd[285938]: Closed D-Bus User Message Bus Socket.
Feb 9 21:03:07 instance-prod-2 systemd[285938]: Reached target Shutdown.
Feb 9 21:03:07 instance-prod-2 systemd[285938]: Starting Exit the Session...
Feb 9 21:03:07 instance-prod-2 systemd[1]: user#1000.service: Succeeded.
Feb 9 21:03:07 instance-prod-2 systemd[1]: Stopped User Manager for UID 1000.
Feb 9 21:03:07 instance-prod-2 systemd[1]: Stopping /run/user/1000 mount wrapper...
Feb 9 21:03:07 instance-prod-2 systemd[1]: Removed slice User Slice of UID 1000.
Feb 9 21:03:07 instance-prod-2 systemd[1]: run-user-1000.mount: Succeeded.
Feb 9 21:03:07 instance-prod-2 systemd[1]: user-runtime-dir#1000.service: Succeeded.
Feb 9 21:03:07 instance-prod-2 systemd[1]: Stopped /run/user/1000 mount wrapper.
Feb 9 21:03:39 instance-prod-2 collectd[1307]: write_gcm: can not take infinite value
Feb 9 21:03:39 instance-prod-2 collectd[1307]: write_gcm: wg_typed_value_create_from_value_t_inline failed for swap/percent/value! Continuing.
Feb 9 21:03:39 instance-prod-2 collectd[1307]: write_gcm: can not take infinite value
Feb 9 21:03:39 instance-prod-2 collectd[1307]: write_gcm: wg_typed_value_create_from_value_t_inline failed for swap/percent/value! Continuing.
Feb 9 21:03:39 instance-prod-2 collectd[1307]: write_gcm: can not take infinite value
Feb 9 21:03:39 instance-prod-2 collectd[1307]: write_gcm: wg_typed_value_create_from_value_t_inline failed for swap/percent/value! Continuing.
Feb 9 21:04:01 instance-prod-2 systemd[1]: Started /run/user/1000 mount wrapper.
Feb 9 21:04:01 instance-prod-2 systemd[1]: Created slice User Slice of UID 1000.
Feb 9 21:04:01 instance-prod-2 systemd[1]: Starting User Manager for UID 1000...
Feb 9 21:04:01 instance-prod-2 systemd[1]: Started Session 19719 of user user_12345.
Feb 9 21:04:01 instance-prod-2 systemd[1]: Started Session 19720 of user user_12345.
Feb 9 21:04:01 instance-prod-2 systemd[1]: Started Session 19721 of user user_12345.
Feb 9 21:04:01 instance-prod-2 systemd[1]: Started Session 19722 of user user_12345.
Feb 9 21:04:01 instance-prod-2 systemd[285996]: Started Mark boot as successful after the user session has run 2 minutes.
Feb 9 21:04:01 instance-prod-2 systemd[285996]: Reached target Paths.
Feb 9 21:04:01 instance-prod-2 systemd[285996]: Starting D-Bus User Message Bus Socket.
Feb 9 21:04:01 instance-prod-2 systemd[285996]: Reached target Timers.
Feb 9 21:04:01 instance-prod-2 systemd[285996]: Listening on D-Bus User Message Bus Socket.
Feb 9 21:04:01 instance-prod-2 systemd[285996]: Reached target Sockets.
Feb 9 21:04:01 instance-prod-2 systemd[285996]: Reached target Basic System.
Feb 9 21:04:01 instance-prod-2 systemd[285996]: Reached target Default.
Feb 9 21:04:01 instance-prod-2 systemd[285996]: Startup finished in 40ms.
Feb 9 21:04:01 instance-prod-2 systemd[1]: Started User Manager for UID 1000.
Feb 9 21:04:06 instance-prod-2 systemd[1]: session-19722.scope: Succeeded.
Feb 9 21:04:07 instance-prod-2 systemd[1]: session-19720.scope: Succeeded.
Feb 9 21:04:09 instance-prod-2 systemd[1]: session-19719.scope: Succeeded.
Feb 9 21:04:10 instance-prod-2 systemd[1]: session-19721.scope: Succeeded.
Feb 9 21:04:10 instance-prod-2 systemd[1]: Stopping User Manager for UID 1000...
Feb 9 21:04:10 instance-prod-2 systemd[285996]: Stopped target Default.
Feb 9 21:04:10 instance-prod-2 systemd[285996]: Stopped target Basic System.
Feb 9 21:04:10 instance-prod-2 systemd[285996]: Stopped target Sockets.
Feb 9 21:04:10 instance-prod-2 systemd[285996]: Stopped target Paths.
Feb 9 21:04:10 instance-prod-2 systemd[285996]: dbus.socket: Succeeded.
Feb 9 21:04:10 instance-prod-2 systemd[285996]: Closed D-Bus User Message Bus Socket.
Feb 9 21:04:10 instance-prod-2 systemd[285996]: Stopped target Timers.
Feb 9 21:04:10 instance-prod-2 systemd[285996]: grub-boot-success.timer: Succeeded.
Feb 9 21:04:10 instance-prod-2 systemd[285996]: Stopped Mark boot as successful after the user session has run 2 minutes.
Feb 9 21:04:10 instance-prod-2 systemd[285996]: Reached target Shutdown.
Feb 9 21:04:10 instance-prod-2 systemd[285996]: Starting Exit the Session...
Feb 9 21:04:11 instance-prod-2 systemd[1]: user#1000.service: Succeeded.
Feb 9 21:04:11 instance-prod-2 systemd[1]: Stopped User Manager for UID 1000.
Feb 9 21:04:11 instance-prod-2 systemd[1]: Stopping /run/user/1000 mount wrapper...
Feb 9 21:04:11 instance-prod-2 systemd[1]: Removed slice User Slice of UID 1000.
Feb 9 21:04:11 instance-prod-2 systemd[1]: run-user-1000.mount: Succeeded.
Feb 9 21:04:11 instance-prod-2 systemd[1]: user-runtime-dir#1000.service: Succeeded.
Feb 9 21:04:11 instance-prod-2 systemd[1]: Stopped /run/user/1000 mount wrapper.
####################### FROZEN #####################
Feb 10 09:29:18 instance-prod-2 kernel: Command line: BOOT_IMAGE=(hd0,gpt2)/boot/vmlinuz-4.18.0-240.10.1.el8_3.x86_64 root=UUID=0d7450f2-b70b-4208-bfe4-8>
Feb 10 09:29:18 instance-prod-2 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers'
Feb 10 09:29:18 instance-prod-2 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers'
Feb 10 09:29:18 instance-prod-2 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers'
Feb 10 09:29:18 instance-prod-2 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256
Feb 10 09:29:18 instance-prod-2 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'standard' format.
Feb 10 09:29:18 instance-prod-2 kernel: BIOS-provided physical RAM map:

App crash report in device logs

I'm facing a problem with a crash after testing my application for more than 20 minutes. It is crashing somewhere but I'm not able to find where it is crashing. I enabled NSZombiaEnable but still, I don't get any correct report.
Below is my device crash log:
Nov 22 23:49:31 unknown UserEventAgent[12] <Notice>: jetsam: kernel termination snapshot being created
Nov 22 23:49:31 unknown com.apple.launchd[1] <Notice>: (UIKitApplication:com.apple.mobilemail[0x56d0]) Exited: Killed: 9
Nov 22 23:49:31 unknown com.apple.launchd[1] <Notice>: (UIKitApplication:com.yqlabsEnterprise.ticketPadPUSHCHAT[0xee8a]) Bug: launchd_core_logic.c:3733 (25562):0
Nov 22 23:49:31 unknown com.apple.debugserver-64[1522] <Warning>: 1 [05f2/1303]: error: ::read ( 4, 0x2ff179d4, 1024 ) => -1 err = Bad file descriptor (0x00000009)
Nov 22 23:49:31 unknown com.apple.launchd[1] <Notice>: (UIKitApplication:com.yqlabsEnterprise.ticketPadPUSHCHAT[0xee8a]) Bug: launchd_core_logic.c:3732 (25562):3
Nov 22 23:49:31 unknown com.apple.launchd[1] <Notice>: (UIKitApplication:com.yqlabsEnterprise.ticketPadPUSHCHAT[0xee8a]) Assuming job exited: <rdar://problem/5020256>: 10: No child processes
Nov 22 23:49:31 unknown com.apple.launchd[1] <Warning>: (UIKitApplication:com.yqlabsEnterprise.ticketPadPUSHCHAT[0xee8a]) Job appears to have crashed: Segmentation fault: 11
Nov 22 23:49:31 unknown SpringBoard[51] <Warning>: Application 'Mail' exited abnormally with signal 9: Killed: 9
Nov 22 23:49:32 unknown SpringBoard[51] <Warning>: Application 'ticketPad' exited abnormally with signal 11: Segmentation fault: 11
Nov 22 23:49:32 unknown kernel[0] <Debug>: launchd[1576] Builtin profile: MobileMail (sandbox)
Nov 22 23:49:33 unknown ReportCrash[1575] <Notice>: Saved crashreport to /Library/Logs/CrashReporter/LowMemory-2012-11-22-234933.plist using uid: 0 gid: 0, synthetic_euid: 0 egid: 0
Please help me.
because it isnt a bad access, there wont be a zombie... its segfaulting.
cause #1:
bad format in an NSLog or stringWithFormat :D (at least for me)
cause #2:
if that's not it try looking for c-code you use that may not exist.
(non-existing symbols)
top hint: cause #3:
the IOS filesystem is case sensitive, maybe the xib name you specify is somehow wrong and it cant load the xib? / a xib itself is corrupt
cause #4:
google for "Assuming job exited: : 10: No child processes" :D

App downloaded from App Store fails to start, no crash or crash log

My App fails to start when downloaded from the App Store, see below. No crash or crash log. Any ideas?
Jul 5 08:54:40 unknown fairplayd.N90[41] <Error>: libMobileGestalt computeUniqueDeviceID: total time for bb to return imei: 0
Jul 5 08:54:40 unknown SpringBoard[52] <Warning>: Unable to obtain a task name port right for pid 1310: (os/kern) failure
Jul 5 08:54:40 unknown com.apple.launchd[1] <Notice>: (UIKitApplication:xxxxxxxx.Xxxxxxx[0x5760]) Exited: Killed: 9
Jul 5 08:54:40 unknown com.apple.launchd[1] <Warning>: (UIKitApplication:xxxxxxxx.Xxxxxxx[0x5760]) Throttling respawn: Will start in 2147483647 seconds
Jul 5 08:54:40 unknown SpringBoard[52] <Warning>: Application 'Xxxxxxx' exited abnormally with signal 9: Killed: 9
Jul 5 08:54:40 unknown kernel[0] <Debug>: AppleFairplayTextCrypterSession::fairplayOpen() failed, error -42110
>
Jul 5 08:55:29 unknown lockdownd[26] <Notice>: 00381000 spawn_and_handle_checkin: Timeout waiting for com.apple.crashreportmover service agent to checkin. spawn=1341471314 select=1341471314 now=1341471329
Jul 5 08:55:29 unknown ReportCrash[1346] <Error>: libMobileGestalt computeUniqueDeviceID: total time for bb to return imei: 0
Jul 5 08:55:29 unknown ReportCrash[1346] <Notice>: Saved crashreport to /Library/Logs/CrashReporter/stacks-2012-07-05-085529.plist using uid: 0 gid: 0, synthetic_euid: 0 egid: 0
Jul 5 08:55:30 unknown lockdownd[26] <Notice>: 00381000 spawn_service_agent: Could not look up service in the map
Jul 5 08:55:30 unknown lockdownd[26] <Notice>: 00381000 spawn_and_handle_checkin: Could not spawn the com.apple.crashreportcopy service agent:InvalidService
>
Jul 5 08:55:30 unknown lockdownd[26] <Notice>: 00381000 spawn_service_agent: Could not look up service in the map
Jul 5 08:55:30 unknown lockdownd[26] <Notice>: 00381000 spawn_and_handle_checkin: Could not spawn the com.apple.crashreportcopy service agent:InvalidService
There's been some problems with Apps downloaded from the AppStore over the last couple of nights being corrupted from the AppStore. Delete and re-install from the AppStore.
http://www.marco.org/2012/07/04/app-store-corrupt-binaries

App Release Version crashes Debug Version works, no idea why

So I have this app I'm working on and after building running the App from XCode on the device everything works fine. However after installing the same App from the App Store (no changes only the provisioning profile is different) the App crashes shortly before it starts.
Console errors:
Fri Sep 30 12:48:42 unknown locationd[540] <Notice>: MS:Notice: Installing: (null) [locationd] (550.32)
Fri Sep 30 12:49:47 itess Ola Portugal[556] <Notice>: MS:Notice: Installing: com.mindovertech.olaportugal [Ola Portugal] (550.32)
Fri Sep 30 12:49:47 itess kernel[0] <Debug>: launchd[556] Builtin profile: container (sandbox)
Fri Sep 30 12:49:47 itess kernel[0] <Debug>: launchd[556] Container: /private/var/mobile/Applications/13E9B45C-84ED-4FD3-BDAA-8527BA34CB3F [69] (sandbox)
Fri Sep 30 12:49:47 itess Ola Portugal[556] <Notice>: MS:Notice: Loading: /Library/MobileSubstrate/DynamicLibraries/Activator.dylib
Fri Sep 30 12:49:47 itess Ola Portugal[556] <Notice>: MS:Notice: Loading: /Library/MobileSubstrate/DynamicLibraries/PDFPatch_CVE-2010-1797.dylib
Fri Sep 30 12:49:49 itess locationd[557] <Notice>: MS:Notice: Installing: (null) [locationd] (550.32)
Fri Sep 30 12:49:51 itess Ola Portugal[556] <Warning>: Warning: Libinfo call to mDNSResponder on main thread
Fri Sep 30 12:49:56 itess ReportCrash[559] <Notice>: Formulating crash report for process Ola Portugal[556]
Fri Sep 30 12:49:57 itess com.apple.launchd[1] (UIKitApplication:com.mindovertech.olaportugal[0xeace][556]) <Warning>: (UIKitApplication:com.mindovertech.olaportugal[0xeace]) Job appears to have crashed: Bus error
Fri Sep 30 12:49:57 itess SpringBoard[28] <Warning>: Application 'Olá Portugal' exited abnormally with signal 10: Bus error
Fri Sep 30 12:49:57 itess ReportCrash[559] <Error>: Saved crashreport to /var/mobile/Library/Logs/CrashReporter/Ola Portugal_2011-09-30-124953_itess.plist using uid: 0 gid: 0, synthetic_euid: 501 egid: 0
Device log:
Incident Identifier: B2ABCA98-7942-4FF2-968A-F5FE4AFDDE4D
CrashReporter Key: fd9745556d91de13e834ad1bbd0bee6c29b17976
Hardware Model: iPod2,1
Process: Ola Portugal [556]
Path: /var/mobile/Applications/13E9B45C-84ED-4FD3-BDAA-8527BA34CB3F/Ola Portugal.app/Ola Portugal
Identifier: Ola Portugal
Version: ??? (???)
Code Type: ARM (Native)
Parent Process: launchd [1]
Date/Time: 2011-09-30 12:49:53.028 +0100
OS Version: iPhone OS 4.0 (8A293)
Report Version: 104
Exception Type: EXC_BAD_ACCESS (SIGBUS)
Exception Codes: KERN_PROTECTION_FAILURE at 0x00816a00
Crashed Thread: 0
Thread 0 Crashed:
0 ??? 0x00816a00 0 + 8481280
(Didn t copy the other threads because they seems to be fine)
Here's one more detail, the device where it crashes is an iPod touch running 4.0
The same App (from the App Store) works perfectly on an iPhone4/iPad2 running 4.3.3
I'm not sure if the problem is with the device or with the OS and what I can do to fix it.
Like I said before If I run the debug version on the device from XCode it works fine.
Any ideas?
Generally this kind of error occurs when we are updating new version on store because of coredata.
If you have changed any thing related to coredata then you need to delete old .Sqlite file at first launch of new version. remember not every time to delete .squlite but only once the application starts and old file exist.
I forgot about this thread. I eventually found out it had to do with iAD, I was using an instruction available on 4.1 but not 4.0

Error on installation on iPhone, iPod devices "Application exited abnormally with signal 11: Segmentation fault"

I tried to install my application on the devices by using Build and Run from xCode, the application works fine, but when I quit, in the springboard there's no icon for the application, I found the following log on the device's console (which I takes from the Organizer application)
Wed Oct 20 17:05:48 iPod-01 com.apple.launchd[1] (UIKitApplication:com.company.app[0x4cbb][1689]) <Notice>: (UIKitApplication:com.company.app[0x4cbb]) Bug: launchd_core_logic.c:3252 (24226):3
Wed Oct 20 17:05:48 iPod-01 com.apple.launchd[1] (UIKitApplication:com.company.app[0x4cbb][1689]) <Notice>: (UIKitApplication:com.company.app[0x4cbb]) Bug: launchd_core_logic.c:2681 (24226):10
Wed Oct 20 17:05:48 iPod-01 com.apple.launchd[1] (UIKitApplication:com.company.app[0x4cbb][1689]) <Notice>: (UIKitApplication:com.company.app[0x4cbb]) Working around 5020256. Assuming the job crashed.
Wed Oct 20 17:05:48 iPod-01 com.apple.launchd[1] (UIKitApplication:com.company.app[0x4cbb][1689]) <Warning>: (UIKitApplication:com.company.app[0x4cbb]) Job appears to have crashed: Segmentation fault
Wed Oct 20 17:05:48 iPod-01 com.apple.debugserver-48[1688] <Warning>: 1 [0698/1403]: error: ::read ( 4, 0x3809f4, 1024 ) => -1 err = Bad file descriptor (0x00000009)
Wed Oct 20 17:05:48 iPod-01 SpringBoard[27] <Warning>: Application 'AppPlus' exited abnormally with signal 11: Segmentation fault
On the simulator, there's no error at all. Someone has any thought about the issue?