Get number of hours worked in Bugzilla using XML-RPC - bugzilla

I'm using Bugzilla 4.0 and can't seem to find a way to get the number of hours worked on a bug by RPC. It still seems like though Bugzilla saves the information in the comments of a bug but even though Additional Hours Worked: x is present when looking inside Bugzilla it is NOT a part of the comment after retrieving it via RPC. I do this with the same user so it shouldn't be a permissions issue. How would I get the hours worked on a bug via RPC?

This seems to be unsupported at this time and I have added a request for it at: https://bugzilla.mozilla.org/show_bug.cgi?id=643681

Related

Error parsing info: Unable to retrieve video metadata

I have made a music bot, and i get the error below everytime I play a song
Error parsing info: Unable to retrieve video metadata
I have googled and found other people who faced the same issue but got it fixed by uninstalling node-ytdl and reinstalling it or by updating to version 3.4.2, i have done both and neither worked.
A few others told that it worked once they started to use a vpn, however even while the bot is hosted on repl.it, i come across the same issue.I do not know how to change server on repl or if its even possible, however if any of you do know how to do so, I would appreciate it if you would tell me.
I am new to coding so please bear with me, thank you.

Azure Machine Learning service (preview) - Minified React Error prevents from viewing experiment

I have not been using Azure Machine Learning service (preview) for too long. To my knowledge it only has been released since the last Microsoft Ignite conference. That's why I think I can not find my question on StackOverflow or any other forum for that matter.
It is as follows:
With help of the Azure Machine Learning service SDK in Python I created my experiment in a Jupyter Notebook locally.
I then configured it to run a Hyper Drive config,
The results came in one by one (as I had a total of 50 runs and 4 simultaneously). It took 7 hours to complete the Hyper Drive run in total.
The next day I went to portal.azure.com to view the results and that worked. I was able to see every run of the hyperdrive and could even compare results.
But then....
I have no clue to as what could have happened, but when I tried to navigate to the experiment again I got a blank white screen. When opening the dev console via F12 I got so see a lot of red errors. All from react. I have zero experience with react, but I am quiet sure that this is the error. React errors when viewing experiment
Hope someone can help. Thanks in advance.
Thanks for reporting this! It's a known issue and a fix is being rolled out.

pgAdmin 4 version 3.0 Query tool Initializing error on opening any query editor in browser

When clicking on a table and then selecting 'Query tool' results in an error.
I also couldn't find the log folder of pgAdmin.
It doesn't matter where I open the editor it always shows this editor.
I have re-installed multiple times, version: 10.4 and 9.6.9 search on stackoverflow for a lot of different solutions like resetting the layout or changing the local ip from 127.0.0.1 to localhost.
My teacher also never saw this before, so he couldn't help me either.
If anyone has had this problem before I would really know what the solution is. Apart from re-installing windows.
EDIT:
I have re-installed windows... And stept away from pgAdmin now. I use another tool (HeidiSQL).
This problem has plagued me for months. Finally, I stumbled across a fix that worked for me. I hope it helps you.
In the Internet Explorer browser that is running pgAdmin4 I turned off SmartScreen Filter under Settings -> Safety, and this instantly fixed the problem.

Google App Engine Error: DNS lookup failed for URL: http://metadata.google.internal

I've been working on a small Google App Engine (standard environment) project that uses Cloud Endpoints v2. My code is largely based on the quickstart provided by Google.
Everything was working fine, but I re-deployed today after having not looked at it for a few weeks, and I'm getting the following error when I attempt to call the endpoint:
error: An error occured while connecting to the server: DNS lookup failed for URL: metadata.google.internal
This wasn't happening before. It seems to be happening when the endpoints package is being imported by Python.
My endpoint doesn't do anything fancy - I haven't changed the source from the sample EchoApi. The error ends up in the GCP Logging console no matter if I try to access the API through the API Explorer or via Curl.
I don't get any errors during deployment.
Edit #1
Some further information:
The error originates from within Google's code that is included with the google-endpoints package which I've included in my lib folder, per
the documentation. Specifically, the error occurs on line 54 of google/api/control/wsgi.py.
Basically, it's making a request to metadata.google.internal using urllib2.
I'm guessing this address is only available from within the Google Cloud, and that for whatever reason, the instance that's hosting my app can't do a DNS lookup on it.
Edit: #2
Dug a bit further.
It seems that the error originates in the google-endpoints-api-management package. Changes committed to that package on October 19th seem to have introduced additional platform reporting. metadata.google.internal is queried to check if the code is running within the Google Container Engine, then it blows up, because the metadata address doesn't resolve.
Here's the commit:
https://github.com/cloudendpoints/endpoints-management-python/commit/0a37d0e443091053ed03e455e06d3a0ae770999f
The google-endpoints package only requires google-endpoints-api-management >=1.0.0b1. On my end, things were working fine on version 1.0.0b2, but then I built a new lib folder, which brought down 1.0.0b5, and things went sideways. Required packages haven't changed between b2 and b5, so I'm thinking I may be able to just downgrade back to b2 for the time being. Haven't tried it yet.
Sent the Google Dev an email. Perhaps he'll chime in with further tips.
Edit: 2016-11-07
Tested downgrading the google-endpoints-api-management package to 1.0.0b2. Seems to be working, kludgy a fix as it is. If you're using the lib folder, the following will scrub the newer error-prone wsgi.py file and put back the older one:
pip install -t lib google-endpoints-api-management==1.0.0b2 --upgrade
Not pretty, but it may just get you back in business.
On a side note, the Google engineer promptly replied saying that he would take a look at this issue soon. With luck, endpoints v2 will eventually come out of beta, 'cause I'm really liking it so far.
This will be fixed in an upcoming patch to the google-endpoints-api-management package (which will be 1.0.0b6). It will probably be released sometime on Monday, 11/6.
If you'd like to continue testing right away and this error is blocking you, you can go back to 1.0.0b4 until 1.0.0b6 comes out. Everything should still work as normal with that version.
Thanks for bringing this to our attention! We're doing our best to iron out all of these wrinkles now during beta in preparation for our first general release.
EDIT: 1.0.0b6 has been released and resolves this issue. Thanks for your patience during our beta phase!
(Posted solution on behalf of the OP).
Google has released version 1.0.0b6 of the google-endpoints-api-management package to address this issue. It solved the problem for me. For anyone who is encountering this problem, clean out your lib folder and re-install the google-endpoints package. This will bring down the new google-endpoints-api-management package with it.
Thanks to Brad at Google for really quick action on this.

GAE downloading log with LogQuery

My application generates many, many log messages. I would like to download only messages in a specific time interval. I looked around and LogQuery seems to be a good choice.
It seems like LogQuery, as specified here:
https://developers.google.com/appengine/docs/java/javadoc/com/google/appengine/api/log/LogQuery.Builder
, let one specify a start and end time in milliseconds.
However, there is no such method withStartTimeMillis or withEndTimeMillis in the GAE library that I downloaded from the web. I'm using GAE SDK 1.7.1. There are only withStartTimeUSec and withEndTimeUsec which deal in microsecond.
What's amiss here?
LogQuery.Builder has both methods: withStartTimeMillis(long startTimeMillis) and withStartTimeUsec(long startTimeUsec).
Also, the current version is 1.8.1, so you should upgrade to that before starting to resolve errors.

Resources