Retrieving recorded images stored in AXIS recorder located in remote area - axis

I am planning to buy and use AXIS Camera Station S2208 Appliance
, and seeking way to retrieve images stored in this recorder in remote area via API. (not retrieve from camera, but recorder)
I guess VAPIX (or ONVIF) API is responsible for this task, but not sure where the exact description is (I looked over VAPIX-library page, but found no clue).
Questions are as follows
To begin with, is it possible to retrieve images from recorder via VAPIX (or ONVIF) ?
If it is possible, where is the description in VAPIX-library page (from Network video, Applications, Audio systems, Physical access control or Radar)?
If not, are there any ways to do it?
I also searched in AXIS Camera Station User Manual, and found out Developer API. However, it was not clear to me about details.
I posted here because I couldn't get answer from official page.
Any help would be great. Thanks!

Related

Setting trainer resistance using Swifty Sensors and Wahoo's cycling power service extension

I'm using the SwiftySensors CocoaPod to connect to a Wahoo Smart Trainer. It's advertising CyclingPowerService and DeviceInformationService. I've been able to get speed and power values without issue. Wahoo apparently extended the CyclingPowerService standard to allow setting resistance via that service instead of the Fitness Machine Control service.
https://github.com/codeinversion/sensors-swift links out to another Github page dealing with that extension, but that link is broken.
My question is: how should I go about setting the trainer's resistance? Wahoo's app can do it, so the machine is equipped for it. This is the only time I need to change the trainer's settings. Otherwise, I'm just reading sent information and SwiftySensors works great.
I've referenced the following post: Writing BLE to Cycling Control Point - Adding Resistance. Someone there said using CyclingPowerService to set resistance was possible without offering any guidance. I'm not very experienced with Bluetooth, so any information would be great!
Thank you Jordan. That was the answer. The broken link I referenced must have been pointing to the following repo: https://github.com/WahooFitness/sensors-swift-trainers
The following instructions assume that you're already able to connect to the trainer to receive data from it, like speed and power, using the SwiftySensors CocoaPod and the CyclingPowerService. Using the repo linked above, I was able to set the resistance to the Wahoo Snap trainer. Note that after you install that new repo, before you start scanning for sensors to connect to, you need to call
CyclingPowerService.WahooTrainer.activate()
From there, you set the resistance with
if let wahooTrainer = cyclingSpeedService.wahooTrainer {
wahooTrainer.setResistanceMode(resistance: 0.5)
}
The resistance is set using percentages. The value for resistance will be a Float, somewhere between 0 and 1.

Can I playback a CAN log file to a CAN input on V2X device?

I am currently working on implementing a Cohda DSRC V2X device that takes a vehicle CAN input. I don't have access to the vehicle and want to simulate the input with a prerecorded CAN log file from it. If possible, we want to playback the CAN log file into the V2X device. I was directed to look into Vector CANoe/CANalyzer products. After looking into their products, documentation, forums, and FAQs, I have not been able to determine if this is possible. So, can this be done, and if so, how?
Perhaps you could use a replay block in Vector CANoe environment if the log file is in Vector supported format and you do not use any secure communication.
For reference please see:
Vector Public Knowledge Area on logging formats, especially the note on the bottom.

Realtime meta-data/ captioning for live streamed audio

How might I achieve adding a track of accurately aligned real-time "additional" data with live-streamed audio? Primarily interested in the browser here, but ideally the solution would be possible with any platform.
The idea is, if I have a live recording from my computer being sent into Icecast via something like DarkIce, I want a listener (who could join a stream at any time) to be able to place some kind of annotation over a few of the samples and allow them to send only the annotation back (for example, using a regular HTTP request). However, this needs a mechanism to align the annotation with the dumped streamed audio at the server side, and in a live stream, the user AFAIK can't actually get the timestamp in the "whole" stream, just from when they joined. But if there was some kind of simultaneously aligned metadata, then perhaps this would be possible.
The problem is, most systems seem to assume you "pre-caption" or multi-plex your data streams beforehand. However, this wouldn't make sense for something being recorded and live-streamed in real-time. Google's examples seem to be mostly around their ability to do "live captioning" which is more about processing audio in real-time then adding slightly delayed captions using speech recognition. This isn't what I'm after. I've looked into various ways data is put into OGG containers, as well as the current captioning like WebVTT, and I am struggling to find examples of this.
I found maybe a hint here: https://github.com/w3c/webvtt/issues/320 and I've been recommended to look for examples by Apple and Google using WebVTT for something along these lines, but cannot find these demos. There's older tech as well (Kate, CMML, Annodex, etc) but none of these are in use and are completely replaced by WebVTT. Perhaps I can achieve something like this web WebRTC, but I'm not sure this gives any guarantees on alignment and it's a slightly different technology stack that I am looking at in this scenario.

How to fetch playback statistics

I'd like to know if it's possible (and if yes: how) to fetch statistics for our soundcloud tracks via the soundcloud API.
I know that i can get the current playback count of my tracks but is it possible to get the playback count of yesterday and so on to draw a custom graph ?
Or do i have to poll the api and record the data myself?
Not officially. Latest and official api does not support this feature yet. It voilates the tos.
Check that question first, before you read further:
soundcloud: Is api-v2 allowed to be used and is there documentation on it?
For experimental use, you can try this:
soundcloud.com/{username}/stats
Is a good starting point to take a deeper look on the networks tab in the console.
Example calls:
totals -> all tracks
https://api-v2.soundcloud.com/users/soundcloud:users:1672444/stats/totals?from=1427846400000&to=1428364800000&client_id=b45b1aa10f1ac2941910a7f0d10f8e28&app_version=1d6991e
plays -> individual tracks
https://api-v2.soundcloud.com/users/soundcloud:users:1672444/tracks/soundcloud:tracks:148976759/stats/timeseries/plays?from=1427846400000&to=1428364800000&resolution=day&client_id=b45b1aa10f1ac2941910a7f0d10f8e28&app_version=1d6991e
The second call will give you, what you want - an array with the daily plays.
To make the calls working i had to manipulate the request headers.
Origin https://soundcloud.com
Referer https://soundcloud.com
Host api-v2.soundcloud.com
Authorization (copied from network tab)

Getting current location and posting it to a web service

Ok... maybe forbidden by Apple, but assuming this is permitted. What would be the best way to get the current location (not difficult) and posting it to a web service (also not much of a challenge). The app is for an artist who wants people to see where he is at the moment. Personally, I wouldn't want people to know where I am 24/7 but that's performance art I guess. An exact location is not wanted but a town or area would be preferred. I could modify the string when it gets to the server and make it more vague by rounding the LOG and LAT.
The app the fans would use would retrieve his current location from the server and update the map view with a location.
I suppose one way would be for him to find out his coordinates and post that to a web page but I'd like it to update automatically.
A response with code is not necessary, just a general opinion on methodology/advisability would be appreciated
I would probably give him a webpage to go to that records his position every time he visits. It would be a private URL, and perhaps use some lightweight authentication. For versatility, I'd record the exact position in the database (in case he changes his mind on how it should work later on). However, you could use an external service or another purchased database with zips or city information with lat/lng data, so you could just provide users of the client app with his general vicinity rather than exact location.
The alternative would be to give him an ad hoc app that does the same thing, but putting together that location recording webpage would be a much simpler process. Could be handy for other things, too.
Here's a little writeup about getting location from mobile Safari, just as a reference: http://mobiforge.com/developing/story/location-iphone-web-apps
Why wouldn't you just use google latitude and read out his status?
I think this should really be done as a web page web-app. You can get coordinates from the device through mobile safari and you can skip the app approval process. You can also set it up so it runs fullscreen on his phone and he won't know the difference. You can also put a password on it.
There is a good chance it would not get approved as a regular app.
You can always meta refresh the page to keep it updating.
http://smithsrus.com/gps-geolocation-in-safari-on-iphone-os-3-0/
"Artist" must log in to his application (thus insuring that he want to share his coordinates with your webserver), then you use CLLocationManager to get his coordinates and make a reverse geocoding using Google Maps API to get country and city for given coordinates.
Can't see anything special about it :)