Difficulty in replicating Ableton Push MIDI CC functions using AudioKit framework - swift

I'm using AudioKit framework to implement MIDI in one of my hobby projects. In this project, I'm trying to make an app which has navigation buttons (left right up down) and play button (Just like Ableton Push MIDI controller has).
To make them function, I first recorded MIDI data that comes out of Push to map all the keys. I then used MIDI Utility by AudioKit as a starter and sent note values from the app to Ableton Live software, where it successfully triggered sounds. (Kept channel as 0)
Now, I'm trying to replicate the cc functionality of arrow keys which are cc54, cc55, cc62, cc63 and cc85 for Play. When I send this cc MIDI data using MIDI Utility, it send midi data successfully to Ableton (as I can see the light feedback), but it simply does not do what Ableton Push hardware controller would've done.
Am I missing something significant?
I also tested that when a button is pressed value goes to 127, and on button release it goes to 0. Despite of replicating it, it still doesn't work.
This problem is not related to AudioKit at all. But someone who has understanding of how midi channels, sending etc works in Ableton Push might be able to help me.

Is ableton identifying your controller as a push? Ableton has special scripting to deal with various controllers (They use Python, and if you hunt around you might find examples). Thats likely both the problem and the solution. The script is not recognizing your software as a push. However it might be possible to create a new device profile in python that'll give you the flexibility to really get in there and tweak.

Related

DualShock 4 Controller Speaker Access

Is there any way to route sound through an individual controller, i.e. play sound A in situation X from controller A and sound B in situation X from controller B?
A questions asking the same thing exists, but it is four years old, so I'm asking again. The person in the old question was using Rewired to access most of the DualShock 4 features, except for the sound, which is not yet possible in Rewired, as far as I have seen in the current documentation. Is there another way to achieve the above?
Unity only supports generic controller functionality, like analog input and button input, but does not have native support for what you want. For specific functionality like audio playing from a DUALSHOCK®4 controller, you'll need the appropriate devkit access and licensing.
If you have a Unity PS4 license, there's built-in Unity support for what you want. As a PS4 developer, I've accessed features like the speaker and lightbar quite easily in the past. Note, however, these features won't work on PC, as there are no drivers for a DUALSHOCK®4 controller that works outside of the PS4 environment.

How to get CC2650 Sensortag to original state

I tried various way's compile and update the firmware on my CC2650 Sensortag, more or less failed that.
Now I want the original functionality back. So I went to where I installed BLE 2.02 (simplelink\ble_sdk_2_02_01_18\examples\hex) and fetched the cc2650stk_sensortag_rel.hex. Programmed that with SmartRF. Now the Sensor tag is back in "near" normal operation. If I detach the programmer, remove the battery, then attach battery. The Simplelink on my Iphone can see a Sensortag 2.0. It notice when I push the buttons, but I can't see anything when using/pushing Sensor view or service explore. What can I do to get it back in it's original working state?
I don't know if I somewhat bricked the device trying to go over to Zigbee OTA and then trying to build/debug the IT resource explorer CC26xx Bluetooth Smart.
This is embarrassing. The SensorTag was working all along, a very important step was missed. Since I already tried uninstall/install the iPhone app I didn't think of switching off/on my bluetooth. The last step fixed the issue. Somehow settings where cashed there.

coding help for physical drum machines

Me and some people are making and selling some physical drum machines, where if you press a button on a board, a related sound is played. Now I'm familiar with Microsoft Small basic and visual and html, however I'm not sure on programming the physical button to play an mp3 audio file, and set it so that you can choose a category of sounds too, thus If you press
Say a number on the machine, then the sounds played will be different. How would I go about doing this? I'm thinking of using a raspberry pi. Also if you could leave an email too, as I may need future help.
You should check out a Chumby Hacker board instead of a raspberry pi as they have audio inputs and inputs included.
This site will tell you all you need to know about the required code for the button.
https://learn.adafruit.com/playing-sounds-and-using-buttons-with-raspberry-pi/overview
While I hour this helps, please keep in mind that this site is for people who have already tried to solve their problem and failed. I produced this answer with a Google search and 5 minutes of reading.

Xcode & Android Radio Scanner

Hi im still not 100% with xcode and android. I have the Droid X and I code with Xcode. My dad is on the Fire Dept. and wants me to code an app for either the android and/or iphone that detects the dispatch's radio signal and receives it and streams it in the background so the firemen could listen to the radio on their phones so they wouldn't have to carry their bulky radios with them everywhere. If there is someone that could help me or point me in the right direction, I would greatly appreciate it.
Thank you, Halokllr
Some bad news Halokllr, the iPhone doesn't have a radio chip, and the FM radio included in some Android models will probably not be able to physically pick up emergency broadcast signals. In order to do what you are suggesting, you would need a web service of some kind that already exists and streams audio for the emergency signals you're interested in. Now there ARE some services like this, though I would strongly suggest that if it is intended to be used for actual first responders that you very carefully research their timeliness guarantees. Some of them may be hours behind live events. Check out http://www.radioreference.com/ as an example service provider.

What does iPhone OS 3.0 need from a programming perspective?

iPhone OS 3.0 is being announced and previewed next week (March 17).
We all know the feature set users want. Copy/paste, MMS, Flash on iPhone, etc.
We'll see about those.
What I'm interested in what does the development community feel the SDK is missing, in need of, to make programming for the platform easier and more productive.
A more complete Interface Builder with support for custom palettes and all sorts of goodies like that.
Better control over the keyboard.
Better unit testing support. (Unit testing can be done, but only on the simulator, and it's very awkward to set up.)
Push notifications. Please.
A more accurate simulator, i.e. one with a more accurate set of frameworks.
The ability to easily build views like the Mail compose window.
For that matter, an in-application compose window.
A better way for apps to share data locally than by invoking URLs.
Access to the calendar, notes, mail (possibly read-only), and bookmarks (again, read-only) databases. Maybe even limited access to the iPod database—even just the ability to read song metadata and access and change the playing song would be helpful.
Some sort of middle ground between UILabel and UIWebView that allows for formatted text without a huge hassle.
More built-in toolbar icons.
The return of the "glass" button style that was in the beta SDK.
A few useful internal views, like UIProgressHUD, exposed.
And last but not least...
A pony.
An easy Javascript bookmarklet installation method for Mobile Safari. (OpenRadar: 1, 2)
UIWebView needs more of UIScrollView's properties and methods, such as contentOffset.
More configurability on some of the built-in behaviors and views, e.g. the button text on UITableViewCell's "Delete" button, or the styles and text of UIAlertSheet/UIAlertView buttons. (Some of these can be done today with undocumented calls, but I'd rather not rely on those.)
More flexibility from UINavigationController, such as the ability to push/pop views that selectively don't display the navigation bar but using the same animations and stack, or more customizability over the navigation bar button labels and behaviors.
The ability to restrict interface orientation per UIViewController, not just accept/reject changes via shouldAutorotate. E.g. I want my main content view to be autorotatable, but I want my navigation hierarchy and settings screens to always display in portrait, even if the content view was rotated to landscape.
libxml and its handy DOM XML parser instead of the SAX-based NSXMLParser.
libcurl w/SSL, or more options and functionality for NSURLConnection.
Ability to check whether a URL scheme is registered. This could be used for apps to detect whether other specific apps are installed, and enable functionality selectively, e.g. when Instapaper detects Tweetie is installed, it can offer a "Post with Tweetie" button. (Disclaimer: That was a plug. I make Instapaper.)
I'm sure I'll think of more, but overall, I'm very happy developing for the iPhone. I'm amazed at the quality and sophistication of the iPhone OS, the SDK, and the development tools given how incredibly young they all are.
I'm surprised no one has mentioned garbage collection yet. Objective-C 2.0 on the Mac supports optional garbage collection. I don't really see any reason it wouldn't work just fine on the iPhone as well and it would eliminate much of the tedium of having to explicitly release objects all over the place.
What I'm hoping most for is to allow iPhones to talk to each other either via Bluetooth or some other means. Granted, they can talk via Bonjour if they are on the same Wi-Fi network, but that's just not convenient enough in 2009. If I'm out with a friend and want to play a multi-player game we first have to find a Starbucks or whatever the heck to get on the same Wi-Fi network. Also, think of the ridiculous amount of social apps you could have if iPhones could talk to each other without needing Wi-Fi. Exchange business cards, flirt with the cute girl over there, etc.
Form a PURE programmers perspective, make XCode as helpful of an IDE as Eclipse or IntelliJ are in the Java world. There's so much time I waste on stupid stuff that the IDE could have found for me as I typed it.
I also don't understand why I can't color buttons without having to use images.
Better multitasking is absolutely key at this point. Android's got it, Palm's WebOS has it - both, it seems, in largely unrestricted and well-implemented fashion. Possibilities:
Push notifications with a good UI (message stack in addition to badging/sound/whatever - if they have to have an extra approval step so apps can't be obnoxious, so be it)
Multiple full processes (not possible with current OS, I realize, but then I've never seen a good explanation why the iPhone doesn't support virtual memory)
Smaller "background" versions of apps that can run in the background - no GUI and a significantly tighter memory constraint
A good mapping API. Let us access the Google Maps abstraction that the Maps application uses !
More Interface Builder goodness
Better simulator
Smart inbox. Incoming messages are routed to installed handlers based on type.
Synchronisation framework that simplifies syncing with desktop & Mobile Me.
Decent landscape support, without the multitude of bugs, especially for the camera picker. Better support for rotation and more control of it.
Access to EXIF data on images from the picker, so we can tell their location
Deeper access to the camera API, so that we are not rail-roaded into the standard photo taker / picker
Push notifications that can launch an application. (In lieu of full multi-tasking, which I don't think we'll get and which could be problematic.)
Better, more intuitive keyboard controls.
API for inter-application messaging.
Access to data from Calendar, iTunes, Mail, Notes and more (with user's permission)
A more accurate simulator, with, for example, ways to limit bandwidth, and use the Mac's camera to actually take a photo.
Phone-phone bluetooth for data exchange
Access to more of the views used by iPhone apps, e.g. the progress HUD, email "blobbing" mechanism for email addresses, thumbnail scrollers, HUD brought up in Photos app, and more.
Less sandboxing. It won't likely happen, but it would always be appreciated for an app to have slightly more power than they currently do (actual filesystem access, for example. even if it was read-only access, it would still allow for more interesting applications to exist).
EDIT: Also, access to the copy/paste API. But I hope that one is obvious to Apple.
My list:
More full-featured IB support as the Mac has
Inter-app Data transfer mechanism (could be C&P, but does not have to be)
Greatly improved camera API with deeper level of control and more flexibility
SDK access to bluetooth and more support for protocols
Real ObjectiveC framework around the address book like the Mac has today.
Warnings similar to the location warning when an app tries to access address book data.
I'm sure whatever they actually have prepared, there will be a few interesting twists.
Ability to send SMS messages without having to have launch the SMS client and have the user type the message.
Access to the raw camera data so that things can be done without having to take a picture and wait for it to save (like you can do with Android)
push notification so that you can launch tasks... would need to be user controllable.
A camera that can focus (I know... have to wait for the next iPhone for that... if they decide to put it in...)
A UIKit level drawing api.
We all know the feature set people want. Copy/Paste, MMS, Flash on iPhone, etc.
I would have thought those specific items were down the SO wish list (although it seems I'm wrong looking at the votes on this comment :-).
MMS is a pretty pointless app when you have eMail. Flash is not an OS issue - Flash could be delivered today.
I don't even want push notifications - they're just a patch, I want background apps. I also want fixes for all the broken APIs like Camera, video and landscape support. Support for CoreImage filters would be nice too but probably too much to wish for.
[[ABAddressBook sharedAddressBook] me] for being able to use the owner's Zip code, phone number, or whatever.
Ability to download files to local storage and sync them back to iTunes or your hard drive
Get EXIF data from photos
Pull all photos at once
Pull all contacts at once
Control screen brightness
Access to music in iPod section
Read access to email and text messages
Access to Safari cookies (so maybe I could make some kind of keep-me-logged-in app.)
fix table view in landscape mode
new camera API with direct access to the camera
distribution code signing automatically when uploading to the app store (instead of code signing in xcode)
ability to request more memory so users don't have to reboot their phones to get rid of background apps
A non-Mac based development envionment.