Silverlight page with multiple page names in google analytics - silverlight

I'm using Google Analytics to track traffic on my Silverlight page.
The same pages in my app are showing up under different names in the top content page.
For example, I have /Home with the most pageviews, then /ClientBin/???.xap/Home with fewer pageviews and then /Default.aspx/Home with even fewer pageviews.
It's the same with other pages (/ManageUsers, /ClientBin/???.xap/Manageusers, /Default.aspx/Manageusers) and so on.
The pageviews are different, so we can't just add them together since we are not sure why this is happening.
I know you can set a default homepage to account for differences in say / and /index.html, but that does't cover all our cases.
What I need to know is why does this happen? Can we just add them together or are some of them a subset of others?
Some of our users are using Out-of-browser, does that count as one of those three pages or is it mixed in with the others?

With some testing using Fiddler I think I've come to a conclusion.
The /ClientBin/???.xap/Home and similar pages are from
Out-of-browser.
The /Default.aspx/Home and similar pages are from
when you run the site from Visual Studio (debugging).
The /Home and similar pages are when the site is accessed on the server.
So in our example we add together the /Home and /ClientBin/???.xap/Home visit numbers to get the real numbers (ignore /Default.aspx/...).

Related

In Wagtail, how to list Pages which are Live and have no privacy setting?

On a Wagtail site, the standard way to get a list of live Pages is something like:
MyPage.objects.live()
However, I want to ensure that I only display pages that are live and have no privacy/visibility setting other than 'Public'. So, to only list pages that every user to the site can see.
What's the best way to do that?
This can be done with
MyPage.objects.live().public()
http://docs.wagtail.io/en/stable/reference/pages/queryset_reference.html#wagtail.core.query.PageQuerySet.public

AngularJS application problems appearance in Google search

I have a personal project which consumes my free time and effort for about a year without significant profit. I have problems with it appearance in Google and would really appreciate to get help here.
This project (http://yuppi.com.ua - similar to craiglist in US) is WEB-based AngularJS 1.2 application that uses PHP rest API hosted on GoDaddy. And in order to make this application popular it have to be very visible in internet and very searchable in Google and users have to be able to share pages via social networks or skype.
According to Google specification, google crawlers doesn't run javascript to get content of a web page before index, so I've added _escaped_fragment_ page that displays content of web page without javascript. For example:
Page: http://yuppi.com.ua/#!/items/sub/18/_
Dirty : yuppi.com.ua/?_escaped_fragment_=/items/sub/18/_
This dirty page will be redirected here where google will see content.
http://yuppi.com.ua/server/crawler_proxy/routee.php?path=/items/sub/18/
So basically I have two versions on HTML file for that page. One version is the one that available to users, which has styles, a lot more HTML tags etc. And the second is the version for Google crawler - very light-weight without any styles. And I am expecting to see clean link to my site in Google, not dirty.
So, If to search all links to a web site in Google you will see that one of the links displays it's "dirty" state.
Another problem is sharing links in Skype.
When I send a link to someone, I am expecting that this link will be transformed to thumbnail image but it is not happens. Instead I see ungly link to my web site.
Please help me to understand how to make happy everyone: users, google crawler, GoDaddy and me.
I was encountering the same problems last year with a big project and we ended to use : https://prerender.io/.
It's a prerendering system that work with a phantomjs browser to detect bot request and render a full html template. It does also instanciate a cache service to not render again a template that haven't change.
Hope it help's.

Why is my angularjs site not completely crawlable?

I have created my first AngularJS website. I have set up pushstate (html5 mode), added fragment metatag, created sitemap in google and tested "google fetch" functionality. After few days, my website is still not completely indexed by google. Google indexed only 1 url instead of 4 (my sitemap contains 4 url's). My website is Tom IT. This main page is index, but this subpage that is also in the sitemap (you can find my sitemap in sitemap.xml in the root of my domain tom-it.be), does not appear in search results. I also added robots.txt.
Google crawlers can parse the pages that generated by SPA and appear at SERPs, but not immediately, may need several days. In my experience, use AngularJS may need 3 days, and use EmberJS need 7 days.
If your website wants to be crawled completely, the important information should put in HTML, or use other techniques, for example, prepare another page for crawlers, server pre-rendering or PhantomJS.

Where to get a web graph with corresponding web pages dataset

I'm trying to implement PageRank algorithm on a set of web pages, for that I need a sample dataset of web pages, and the web graph corresponding to them, this web graph represents the links between the pages that the data set contains.
I need the web graph so I can get the transition matrix and do the calculation needed. Example:
URL1 -> URL2
URL3390 -> URL5
URLxxxx is an id, somehow mapped to the corresponding web page
My question is: how/where can I get this resource (I've tried many links on the internet but nothing really helps), I would also like it to be not of a very large size, (internet connection limitation), if I can't have this as it is, could sou give me some advice on what I should do?
Update: for people who may consider this off topic, and they may be right, networks like Software Recommendation or on Computer Science, don't even have corresponding tags, and doesn't really fit the kind of this question, I appreciate your help.
May be Site Visualizer is the tool you're looking for. The app has the feature to generate visual sitemap.
Download and install the app (Standard or Pro version), click Create new project toolbutton, type the URL of the website you need to crawl, and then click Start button.
After the crawling is finished, click Draw button on the Visual Sitemap tab. Graph of the website will be drawn as a set of pages (rectangles) and links (lines with arrows). Click on a box to select the certain page and highlight its outbound links:
Dataset of all links of the website you can get by using All Links report (on the Reports tab). 'From URL' and 'To URL' columns are what you need.
Besides of that, you can represent a dataset of pages or links of the crawled website by using your particular SQL query. For instance, go to the Database tab, type the following query and click Execute toolbutton:
SELECT * FROM links WHERE link_type='A'
The resultset will contain only A-tag links, excluding images, CSS files, JS, etc.
The program has full-featured 30-days trial period, so you can carry out your tasks for free.
you might try searching for datasets used in supplementary information for PageRank papers. Here's an example:
this paper: http://langvillea.people.cofc.edu/ReorderingPageRank.pdf
uses this dataset:
http://www.cs.cornell.edu/Courses/cs685/2002fa/data/gr0.California
which supposedly contains 9,664 nodes and 16,773 links. The links are at the end of the file and appear to be in a connection format similar to what you're looking for.
from this page (which also has other datasets):
http://www.cs.cornell.edu/Courses/cs685/2002fa/
here's a few other pages that aggregate network datasets:
http://snap.stanford.edu/data/, see particularly
http://snap.stanford.edu/data/web-Stanford.html
http://www.datawrangling.com/some-datasets-available-on-the-web
http://networkdata.ics.uci.edu/resources.php
good luck!

Hashtag urls in mobile devices? How do they work?

Ok, probably a stupid question. But here goes... I know on regular websites a hashtag is an anchor to an element on the page, however, how does this work in mobile devices? It seems to me that mobile devices can use hashtags to slide to another section on the same page, sort of like another page.
How do urls work like that in mobile devices?? I've been to the jquerymobile.com site and noticed the webapp builder on the homepage, offers the ability to change features to slide, slideup, slidedown, fade, pop, etc. etc., however, the only URL to choose from is always "HOME" for these features. And home points to #page1, which is odd that it has a hashtag in it. How do mobile devices work with these hashtags? More importantly, how do I add the ability to slide to that hashtag, as in the way jquerymobile.com implies in the webapp builder?
I've tried doing my own research on this via google, but no luck, as mobile hashtags keeps bringing up other results, not what I am asking for here.
There's no difference between handling of the hash parts of URL in desktop vs. mobile browsers. The key point is that Javascript has access to the URL, including the hash part, and can do any programmatic thing based on the contents of the URL. So if the page author wants to slide in another section in response to changing the hash part of the URL, they can do that. People do the same thing on "desktop" sites, too.

Resources