I'm looking at scraping values off a Cordex rectifier system's web page. I have maybe a couple 100 of them to do and want to put the data all in a nice live table so when there are power outages we can monitor the battery voltages of all our buildings. We have a tool for that but we will be losing it soon. They are moving the alarms to another system that is inferior and does not show voltages. These pages require a login, uses https (making it hard to use packet sniffer to analyze it) and is also very heavy on JS, so looking at just the html you don't get any values.
I assume there must be some python libraries or something that could help me do that, and maybe some firefox extensions that can let me analyze it more. From using the "network" tool in Firefox I've found some basic stuff like json files I can get data off of, except when I try to do it manually it does not have the context of a login, so I presume my script would need to perform a login first, keep cookie info etc... Basically lot of digging around I'll need to do to get everything right so the session data is kept.
I'm looking mostly for guidance on how this sort of thing would normally be done. Am I more or less going down the right path looking at the "network" feature of Firefox and just analyzing the GET/POST requests and headers etc to see how the data is fetched, or is there perhaps an easier way? I would more than likely write the scraper in python and have the values go into a clean easy to read file, then I would use a monitoring software to simply monitor the file data points. Those details are minor and I'm not too worried about that part. It's mostly getting the data.
I assume there must be some python libraries or something that could help me do that, and maybe some firefox extensions that can let me analyze it more. From using the "network" tool in Firefox I've found some basic stuff like json files I can get data off of, except when I try to do it manually it does not have the context of a login, so I presume my script would need to perform a login first, keep cookie info etc... Basically lot of digging around I'll need to do to get everything right so the session data is kept.
I'm looking mostly for guidance on how this sort of thing would normally be done. Am I more or less going down the right path looking at the "network" feature of Firefox and just analyzing the GET/POST requests and headers etc to see how the data is fetched, or is there perhaps an easier way? I would more than likely write the scraper in python and have the values go into a clean easy to read file, then I would use a monitoring software to simply monitor the file data points. Those details are minor and I'm not too worried about that part. It's mostly getting the data.