Get resource consumption details and costing of Azure VM using REST - rest

I am in need to get current resource consumption details of specific VM and the costing of that VM according to consumption using REST api.
[Want Azure API to which I can pass the VM name/id and result would be Usage details and cost. They do have usage and Rate card API End Points but again it needs taking one Val from Usage and one relative from Rate multiplying it to get total]

Is that what you are looking for?
The Azure billing API allows to get data of your usage and the money
that you have to pay for your resources.
https://codehollow.com/2017/02/using-the-azure-billing-api-to-calculate-the-costs/
Azure Billing REST API Reference:
https://msdn.microsoft.com/en-us/library/azure/mt218998.aspx

Related

How to get sample data from sales force developer account through rest api

I need to get sample data from sales force developer account but i am unable to find any API to get any data from sales force.
Is there any API exposed by sales force to accomplish the task.
Salesforce exposes a large number of thoroughly documented APIs.
Access to record data is most commonly achieved through the standard REST API.
You may be most interested in the endpoints named under Working with Records, which provide access to record data, or Working with Searches and Queries.

SharePoint throttling

We have a lot of users that use SharePoint and also we use only application token to manage the data. So we send lots of requests and we are facing throttling. So the questions are:
Are there any possibilities to increase the requests limit?
How can I see my current quota/resource usage?
You can not request a limit increase, and Microsoft does not provide any specific information regarding the algorithm they use to decide when or why to begin throttling.
The best option you have is to apply incremental back-off logic to account for throttled calls:
https://learn.microsoft.com/en-us/sharepoint/dev/solution-guidance/handle-sharepoint-online-throttling-by-using-exponential-back-off
You have to contact Microsoft support and open a ticket in order to get information regarding which service principals and users are being throttled. An engineer will have to pull the info from their database.
They can not give you specific quota/resource information. The will only tell you how many calls were throttled for a given principal.

Bluemix alchemy api cost

I am confused about the cost of the alchemy api. I am trying to figure out the cost of 3000 requests but it looks as if each request costs $0.007 which seems rather expensive. or is it $0.007 per 1000 requests?
You are referring to the standard pay per use plan. Basically the main concept is the API Event, not the request.
As stated in Alchemy API Service terms:
Events are calculated for and included in the information returned by each API call. Calculations vary by the type of service:
Events related to the use of the IBM AlchemyLanguage API will be calculated based on the number of enrichments returned by the API. Many of the API calls require 1 event; some API calls require multiple events.
Events related to the use of the IBM AlchemyVision API will be calculated based on the number of enrichments returned by the API. Many of the API calls require 4 event; some API calls require multiple events.
Events related to the use of the IBM AlchemyData News API will be calculated based on the following facets:
a) the number of enrichments returned by the API; b) the volume of data returned by the API; and c) the length of time, as specified in the API call, used to determine the number of news objects included for analysis.
In the Free plan the customer is entitled to a free allotment of 1.000 API Events per day across all services. In the Standard plan you will be charged per API Event.
The Standard plan is convenient when you need many concurrent calls.
If you are not deploying a large-scale application maybe using some cache services in the right way you can work in the free tier.
Take a look at Pricing for AlchemyAPI Services to read a detailed description of what is a transaction and how AlchemyAPI define those.

Increase Batch Quota in Google Core Reporting API

Does anyone know if there is a way to increase the quota limit of 10 queries when batching calls to the core reporting API?
This question/answer mentions the limit of 10: How can I combine/speed up multiple API calls to improve performance?
If I try to add more than 10 queries to the batch only the first ten are processed, each one after that contains a 403 quota exceeded error.
Is there a pay option? Would love to speed up the process of reporting on GA data for a bunch of URLs. I looked in my Google Developer's Console under the Analytics API where there is an option to increase the per-user limit and a link to request additional quota but I don't need total quota to increase, only allowed batch requests.
Thanks!
Quota is the number of requests you are allowed to make to a Google API without requesting permission to access more. Most of the Google APIs have a free quota, a number of requests Google lets you make without asking for permission to make more request. There are project based quotas and user based quotas.
Unless it says other wise APIs Quotas are projects based not user based.
User quota example
Per-user limit 10 requests/second/user
Some Quotas are user based, a user is normally the person that has authenticated the request. Every request sent to google contains information about who is making the request in the form of the IP address where the request came from. If you have your code running on a server the IP address is the same all the time so Google sees it as the same user. You can get around his by adding a random Quotauser to your request this will identify the request based upon different users.
If you send to many requests to fast from the same user you will see the following error.
userRateLimitExceeded The request failed because a per-user rate limit
has been reached.
The best way to get around this is to use QuotaUser in all of your requests, and identify different users to Google. Or just send a random number every time should also work.
Answer: You can't apply for an extension of the flood protection user rate limit. But you can get around it by using QuotaUser.
more info on quotas can be found on Google developers console APIs

Google Cloud Storage free trial quota details

Is there a possibility to set a quota (usage limit) for Google Cloud Storage while I am in the free trial quota? Otherwise I have unlimited risk until I notice it and set the quotas, when the free trial quota ends.
(I can set quotas on the quotas tab of my google api console for activated apis like Prediction API.)
"Google Cloud Storage offers a free trial quota until December 31, 2012".
Will there be any free trial quota afterwards?
Google Cloud Storage does not implement usage limits on the XML API or for HTTP reads of public objects. The JSON API has a daily cap ("Courtesy limit") listed in the Google API Console -> Services (https://code.google.com/apis/console/). You can request a change to that cap via the link next to it, but it won't affect the other two options for accessing the data.
The Free Trial Quota is independent of this, it makes the initial usage each month free of charge. Details here: https://developers.google.com/storage/docs/pricingandterms