Has anyone worked with www.atript.com? - affiliate

I want to use atript.com for further monetization of my page. Has anyone worked with it yet?
Is it secure and how long does the integration take and how can I see whether the container is working?
Thx!

I integrated atript already on a couple of websites. I never really had security issues. The integration is super simple, usually took me like 5 minutes. Wherever I could I tried installing it using the Google tag manager.
To check whether the container is working just open your browser's developer console. If you see “loading tripts…” in the console the atript.com container is active and working.

Related

Confusion about Lighthouse report of unused JS for react devtools chrome extension

When I run Lighthouse for the project I am working on, I get
chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/react_devtools_backend.js
as no1 causer of unused JS.
How is that making sense from a performance optimization perspective?
I can't control what extensions the user runs in his browser.
I also can't find anything about the React dev tools in the project.
Would highly appreciate if someone could shed some light on that issue.
I found the issue is because the file of the devTools is installed when the page is loading... and it is useless in production mode.
So go to the extension and right click on it.
Then go to the manage extensions.
Then find the "Allow this extension to read and change all your data on websites you visit" change it to "on click".

How do I make selenium see the network requests made by a web browser?

I have a dotnet Selenium web driver app.
When I'm testing the page one of the things I need to confirm is that a flash object on the page has pulled correct content from a content store on my site. (i.e. the flash object should be loading content from /stuff/info.txt and including that content within the animation.)
As a human looking at this I can use the chrome network tab and see that /stuff/info.txt has been accessed.
How can I make Selenium execute a similar watch and see the network requests made by a web browser?
I did not wrote this, neither tested it however someone did it here: http://www.softwareishard.com/blog/firebug/automate-page-load-performance-testing-with-firebug-and-selenium/
Basically all the requests are exported via netexport and firebug plugins inside a HAR (Http ARchive file)
Please give us your feedback if you give it a try!
Cheers !
I assume you want to automate the process which the developer tools of browsers does. Something like firebug but for verification using Code.
I don't believe Selenium has such features. For now, you will not be able to achieve this.

Need guidance in debugging this DNN issue

I'm a beginner with DotNetNuke and am having trouble debugging this problem. We are using DNN version 5.
On our local development machines our DNN site renders just fine. But we recently deployed to our test server and on there the site only renders fine when logged in as the Host. When viewing the site as just a normal visitor, all the styles are messed up -- it is just a white page with images gone and menus stacked vertically.
Clues / Notes:
We recently added a new .ascx and module
There are javascript errors, but it hard to tell what is causing them
You may ask about the Log Viewer screen. I looked there and I see exceptions, but I think the exceptions there are unrelated ( will look into this)
Works fine in our local dev environment
In server, only renders fine when logged on as Host
What are some steps I can take to try to debug this? Where else can I look besides the Event Viewer to see what is going on here?
Thanks!
What is the "new .ascx" that you deployed? Is that a skin page?
If something is wrong in the skin page and it is the default site page (under Admin > Site Settings), that would explain why non-admin users have issues, but Admin/Host users (where the site's Edit skin is working) don't have issues.
This solved it:
https://stackoverflow.com/a/3831419/614263
Due to permission settings it looks like IIS was blocking requests from the root to javascripts and css files. When looking at the Network traffic in FireFox I noticed that 404 errors were coming from "login.aspx?ReturnURL=../../blahblah.css"
Thanks for all your help and suggestions.

Browser Automation with Selenium: Fingerprints, recognizability and traceability?

I want to use selenium/webdriver to simulate a browser and scrape some website-content with it. Even if its not the fastest method, for me it has many advantages such as executing scripts etc.
For many websites it is forbidden to access them via an automated method, for example search engines like google or bing.
For one tool i need to scrape the estimated resultstat from google for several keywords. This will look like the following: simulate the browser that visits google.com and types in a keyword and scrapes the results, then after a little pause type in the next keyword, scrape the results and so on...
My question is: Is it possible for a website to recognize that I'm using selenium to simulate the browser instead of using the browser by hand? Especially the google case gives me some doubts. I know selenium is partly developed by google or at least by some guys working for google. So does leave selenium some fingerprints or isn't it possible to decide if I'm using the browser by myself or simulated by selenium, even for google?
No, nobody can actually see that you're using Selenium and not hand-operating the browser yourself with WebDriver. I'm not sure about the old Selenium RC, but it should be the same way. Here's how it works:
Selenium opens up a browser with a clean profile (or with a profile you selected)
Selenium is hooked up to the browser so it can steer it, control it. But the browser still does most of the work. Basically, Selenium replaces the user inputs to the browser, but not more.
You can easily verify this by reading the contents of the HTTP headers sent by your browser.
If you ever actually needed Selenium to be recognized by your server, you can use Browsermob-proxy and add a custom header to your requests.
All that said, there is one thing you must be aware of. While there's no way to detect Selenium directly, there can be some indirect clues picked up by the website you're visiting. Those usually include scanning for too many requests made in virtually no time - this might be an issue for you. Make sure your Selenium is behaving like a user.
EDIT 2016/04:
Apparanetly it is possible as https://stackoverflow.com/a/33403473/2930045 states that a company can do it. My guess - and it is nothing but a guess - is that they can run some JS that Selenium installs into the browser to operate.
Signs point to yes, sites are able to regonize that you are using Selenium.
Counter Example: www.stubhub.com detects and blocks my browser instance launched using Selenium while "normal" browsing done manually (not using the browser launched by the Selenium web driver) work with out issue.
See this stackoverflow question for additional details
Can a website detect when you are using selenium with chromedriver?

Checkboxes disappearing when uploading to Google App Engine

I have written a controller to take input from the form containing a list of options. The options have to be selected using checkboxes. I have successfully tested the code on my local machine. But when uploaded to the app engine, the webpage is not displaying the checkbox fields. It is not displaying the options either. I tested with 3 browsers - IE, Mozilla and Chrome and this happens everywhere. But, everything is working on fine on the browsers when running on local machine. I am quite confused as to what is going wrong.
Please help me fix this issue.
Thanks in advance...
You likely need to setup static file/directory mappings in app.yaml.
You need to provide us with more details to help you debug issues. For example, if the browser doesn't display what you expect tell us what it does display. What happens if you view source? Is your HTML there?

Resources