How to export data from Chrome developer tool? - google-chrome-devtools

Network analysis by Chrome when page loads
I would like to export this data to Microsoft Excel so that I will have a list of similar data when loaded at different times. Loading a page one time doesn't really tell me much especially if I want to compare pages.

if you right click on any of the rows you can export the item or the entire data set as HAR which appears to be a JSON format.
It shouldn't be terribly difficult to script up something to transform that to a csv if you really need it in excel, but if you're already scripting you might as well just use the script to ask your questions of the data.
If anyone knows how to drive the "load page, export data" part of the process from the command line I'd be quite interested in hearing how

from Chrome 76, you have Import/Export buttons.

I was trying to copy the size data measured from Chrome Network and stumbled on this post. I just found an easier way to "export" the data out to excel which is to copy the table and paste to excel.
The trick is click Control + A (select all) and once the entire table will be highlighted, paste it to Microsoft Excel. The only issue is if there are too many fields, not all rows are copied and you might have to copy and paste several times.
UPDATED: I found that copying the data only works when I turn off the filter options (the funnel-looking button above the table). – bendur

Right-click and export as HAR, then view it using Jan Odvarko's HAR Viewer
This helps in visualising the already captured HAR logs.

I came across the same problem, and found that easier way is to undock the developer tool's video to a separate window! (Using the right hand top corner toolbar button of developer tools window)
and in the new window , simply say select all and copy and paste to excel!!

In Chrome, in the Developer Tools, under Network, in the Name column, right-click and select "Save as HAR with content". Then open a new tab, go to https://toolbox.googleapps.com/apps/har_analyzer/ and open the saved HAR file.

Note that ≪Copy all as HAR≫ does not contain response body.
You can get response body via ≪Save as HAR with Content≫, but it breaks if you have any more than a trivial amount of logs (I tried once with only 8k requests and it doesn't work.) To solve this, you can script an output yourself using _request.contentData().
When there's too many logs, even _request.contentData() and ≪Copy response≫ would fail, hopefully they would fix this problem. Until then, inspecting any more than a trivial amount of network logs cannot be properly done with Chrome Network Inspector and its best to use another tool.

You can use fiddler web debugger to import the HAR and then it is very easy from their on... Ctrl+A (select all) then Ctrl+c (copy summary) then paste in excel and have fun

I don't see an export or save as option.
I filtered out all the unwanted requests using -.css -.js -.woff then right clicked on one of the requests then Copy > Copy all as HAR
Then pasted the content into a text editor and saved it.

I had same issue for which I came here. With some trials, I figured out for copying multiple pages of chrome data as in the question I zoomed out till I got all the data in one page, that is, without scroll, with very small font size. Now copy and paste that in excel which copies all the records and in normal font.
This is good for few pages of data I think.

In more modern versions of Chrome you can just drag a .har file into the network tab of Chrome Dev Tools to load it.

To get this in excel or csv format- right click the folder and select "copy response"- paste to excel and use text to columns.

You can try use Haiphen, which is a chrome extension that allows you to analyze network traffic and what API calls a web application is making.

Related

Powershell GUI displaying text file

I have an app I developed and made a GUI for. In one of the directories I have a log.txt file.
I have a strip menu option meant to display the log file to the user. I'm wondering what would be the best method for displaying this log file?
Right now I'm just using
notepad log.txt
Which technically works but doesn't seem to be the most efficient or professional way.
I'm guessing it is probably just opening another form with a large read only text box but on the off chance someone has something that works even better I'll appreciate the suggestion.
Info displayed in log is:
Version
Description
Date

TinyMCE and Dynamics365

I have a TinyMCE editor working with Dynamics365. In it, I'm able to get images to copy and paste into the window. However, I'm wondering where these images get posted to and/or stored and what I need to do to bring them back up when the page is closed and reopened.
When I reopen the page I get a list of errors saying the images can't be found (makes sense).
Thanks
When you copy and paste the image into TinyMCE what actual HTML are you getting before you save the content? What plugins are you loading in TinyMCE to help facilitate the copy/paste? There is a free paste plugin and a commercial powerpaste plugin - these would typically result in slightly different HTML depending on the source of what you are copying.
Knowing what sort of HTML you get after the paste would be the first step in determining possible solutions.
We are using URLs for inserting the images into the page.

Searching inside JSONs in Chrome devtools

Is there a possibility to searching inside all JSON objects from all available responses in the network tab? Currently it works, but very randomly and isn't much reliable. Sometimes and especially in a smaller responses it's ok but when you have more assets almost always looking for, e.g. specific params value ends unsuccessfully. Do you know any smart solution of that issue? I've checked and first question associated with it has already few years and Google devs still haven't responded.
Example: I have object ID in response body, but cannot find it by search CTRL+F
I think one way is to save all the response in a file (manually or automatically, if possibile by using a browser extension).
After you have stored all the responses in a file you can parse the file and find things inside the file by using a script or just regex.
You can save the answers (as HAR file) manually (I use firefox) by right clicking on a network response inside the developer console panel.
I found that is the same for chrome.
Look here:
https://developers.google.com/web/tools/chrome-devtools/network/reference
I didn't search if there is a way to automatically store all the responses received by a browser. I'm not sure, but I think it isn't possible :/

Reading Data from excel sheets in Katalon studio

How to read data from excel sheets(sheet 1,sheet 2...etc) in Katalon Studio.
I am able reading the data from one sheet.But I am facing the difficulty with multiple sheets.
Without knowing exactly what the problem you face is, here are my first suggestions to check:
You do have to create different Test Data files for each sheet in your file that you want to use. Two Test Data files for the same spreadsheet file
In your new Test Data inside Katalon, just select your file and then you can select the different sheet. Sheet selection options
If you can't see a sheet in the drop down on your Test Data page (perhaps you just added it to the spreadsheet) refresh your project, or at the very least refresh the Test Data. (ctrl-F5 to refresh the entire project, highlight the Test Object and then right-click and select or hit F5 to refresh the Test Data)
Hope this helps.
You cannot switch excel sheets from Katalon, at least natively. Maybe with some custom programming.

Automation of clicks on webpage using Matlab

So I'm looking for a way to systematically access data from a website. This data is updated every 15 minutes or so, and is generated through a datamart system that makes custom reports following several input parameters: the desired date interval, the specific dataset.
All these parameters require me to click on some specific buttons; I was wondering if it would be possible to automate these click inputs using Matlab (or something else if need be), to retrieve the data and treat it automatically.
Thanks in advance!
I suggest you take a look at http://www.autohotkey.com/. This is a great tool which allows for the automatic clicking on any window (including a browser page) under Windows. It even will let you "search" your screen for pixel images and then click on those images. This would allow you to make a very small bmp file of the link you would like to click on, and then your script can search your page and click directly on the link.
As far as getting data into matlab I'm not exactly sure of the best way to do this, but you might consider saving the html of the page, and then parsing that from matlab.