Hi guys I've a big question.
If I have an app that could be protected by Touch ID protection, how do I have to manage the 3D Touch quick action that interests data? I've to ask fingerprint before the execution of the action, or I let the action be executed without any lock, or maybe I create an alert that tell the user 'Impossible to use quick action while Touch ID protection is actived' ?
So, I ask you the best way to manage this type of problem..code will be well accepted :)
It should be noted that
sensitive information are not to be shown when the app icon is 3D
touched.
What should happen is that you should allow the user to interact with
the 3D Touch but when the app opens the screen based on user's
selection, the app should prompt for Touch ID or passcode alert.
Unless it's verified app shouldn't function as intended since the user
is not authenticated at that point.
In case you want to stop the actions from appearing all together, then it could be done, you would
just need to change the 3D Touch actions dynamically.
Refer to this for details, there are two ways to generate action 3D Touch action items, static and dynamic. Using the dynamic one, you could achieve it.
Related
Update:
void Awake() {
if(UnityEngine.VR.VRSettings.loadedDeviceName.Equals("HoloLens"))
UnityEngine.VR.InputTracking.Recenter();
}
Doesn't appear to do anything. Found this related post titled, "Reset Hololens Origin" inquiring about a similar behavior. Two individuals affiliated with Unity in someway (both bare the 'Unity technologies' profile tag) brought up InputTracking.Recenter(). One adds, "[S]ome changes landed recently, and part of those changes were hooking up InputTracking.Recenter properly on WindowsMR."
The snippet of code I include above is an attempt to implement this behavior. I most definitely could be trying to call this code in an improper/incorrect location but, as it is now, I do not see it working.
end update
When a user opens a Unity Hololens application they are given control over where the application starts by giving them a generic white box to denote the application and allowed to place it any where via the air tap gesture.
Problem:
If a user blooms out of the application but does not close this white box, the application is left in a suspended or 'tombstoned' state. When the application is opened again, the Hololens remembers the old white box placement and reloads the application their.
This isn't desired behavior as with multiple users and or a variety of locations, the application appears to not be loading to the user when the Hololens is bringing the suspended application back into the running state but not updating its position to current user gaze.
Question:
Can one detect, in awake, where the user is looking and move the application manually?
UnityEngine.WSA.Application.windowActivated : Fired when application window is activated.
Somewhere like this?
Any google searches on this question or variations in wording turn up nothing. Unsure of the possibility of moving the app itself.
If this isn't possible would the next alternative be to force the app to close when I want, say for example, in the suspended state? In an attempt to prevent this type of multiple apps open situation.
Basically, we either have remote access to the iPhone or the phone is connected to a network where we can control the phone (send it messages etc.) How can I simulate a swipe without touching the actual phone? I know there are Swipe Recognizers, but I haven't found a way to HARDCODE coordinates to simulate a swipe; for example, without touching the phone, perform the swipe to unlock.
A swipe is input. You'd normally recognize the swipe, either with a gesture recognizer or by handling the touch directly, and then perform some sort of action. If you want to simulate a swipe, just perform the action that would be performed if the user made the equivalent gesture.
For example, if a swipe would normally switch to a different view, simply call the method that switches to that view. If possible, do it with animation so that the user has some visual indication of what's going on.
I'm not entirely sure that what you are trying to do is possible, unless maybe with a script in the Automation instrument. However, if the iPhone is jailbroken, you could install Veency and connect the phone through any VNC client and interact with it that way.
Sorry, I am new to iPhone development and my google searches have failed me, so I have had to resort to posting a potentially idiotic question on SO.
I would like to write an app that, when suspended, performs an action when the user presses the home or power button (presumably to unlock the phone). I read the section in the iOS App Programming Guide's App States and Multitasking and the "Processing Queued Notifications at Wakeup Time" discusses handling queued events upon waking up. However, this isn't what I'm after.
I'd like to know if it is possible to:
From the phone sleeping state (I couldn't find a document for phone states, so I am talking about the case where the user presses the power button to turn off the screen), I would like my app to be ready to respond to the event where the user presses the power or home button (to unlock the phone)
I would like my app to respond to the event that occurs when the user unlocks the phone. I found an answer here that's close.
I don't want any funkiness when calls come in, get ignored, etc. :)
From what I can tell, it's a very gray area in the API around locking and unlocking, and I'd like to verify whether or not I'm wasting my time trying to do this.
It looks like I can use the accelerometer to detect when the phone is locked, but I also assume that I won't be able to count on this behavior in all future versions of iOS.
EDIT - I think I can handle the locking and unlocking requirements by assuming that the application has to be running at the time the phone is locked and unlocked, but I still cannot figure out if it is possible to determine #1 above, which is that the power button has been pressed and the unlock screen is displayed. Likewise, I'd like to know if the power button has been pressed again and is no longer displayed (i.e. screen is off).
All of the behaviors for your app being suspended/backgrounded are in the document you referenced. The AppDelegate will receive these messages and pass them on to whatever view you want to listen for them.
-applicationWillTerminate
-applicationWillEnterBackground
-applicationDidBecomeActive
-applicationWillResignActive
You can set up a notification observer in your view if you would like the view to be notified of any of these events. Then just override them or set up a custom method to do any work that needs to be done.
While the Apple docs might be a little foggy at first, all the information is in there to let you know which state your app will go to and how it will be handled. It is up to you to figure out what your app needs to do for each of these events. Hope this helps.
I am writing an API in C++ Builder that collects information for events on the touchpad of a windows laptop.
This is how I was doing it.
I was creating a window
when the touch pad is touched, I simply paint that information on that window in WM_PAINT event.
But now I dont want to create that window (form), i want to catch all the events, even if user is on desktop screen or on another application's window. If an application that is using my API is running in background i want to be able to get that touch even information in the code. How can I do that??
I hope you are getting my point ... actually i want to do it in a seamless way, otherwise that white form window will irritate the user.
I also want to save these events in a link list and want to return that out of the API is it possible??
I will be very thankful for your time. I really need to work it out in next few hours.
The touchpad is just a mouse like any other. It generates standard mouse events. Use a global WH_MOUSE hook via SetWindowsHookEx() to capture mouse events globally. To replay them, use mouse_event(). Alternatively, use WH_JOURNALRECORD and WH_JOURNALPLAYBACK hooks instead for capture and playback, respectively.
Is it possible to override the default behavior (close app) of the Home Button? And if how do you do that?
well, there's a reasonable alt to the need to detect if the Home button was tapped. i'm currently working on just such a scenario.
my app allows user input (into UITextViews). a customer has asked for an "autosave" feature where, should they (accidentally) tap the Home button during data entry, could i still capture the data entered and save it.
my testing of the sequence of viewWillDisappear in the view that is collecting the data and applicationWillTerminate show viewWillDisappear getting called before applicationWillTerminate. now i fully appreciate that this is the architected sequence Apple intended, BUT this sequence also precludes setting a simple Boolean in applicationWillTerminate to let the particular viewWillDisappear know that it is disappearing specifically due to applicationWillTerminate.
so the issue still stands. how can a UIView detect that it is disappearing specifically as a result of the app terminating?
Even if it is possible, your app will probably not be approved by Apple.
Are you planning this feature for your own private App or do you want to distribute it later on?
The home button is probably the most critical button on the device on not a good idea to change