Posted by Julian Rocco

November 19, 2012

11:29 am

Leave Reply

Web Analytics 101

These days practicing web analytics is not new to most businesses. Web analytics has its roots in the early days of the web (the mid 1990’s) and has grown steadily since then. I, on the other hand, am somewhat of a newcomer to analytics, having recently joined a very talented team of consultants here at Unilytics Corporation. Some of the members on my team have been practicing web analytics and data analysis for over 20 years. That knowledge gathering on best practices and analytics strategy is being passed on to me every day! So for those entering the web analytic space or trying to make sense of web traffic data, this post shares with you what I’ve learned travelling that same path.

A Brief History of Web Analytics:

Since the mid-late 90s IT departments have been monitoring web servers for performance problems and analyzing server log files for page errors.  Then in the early 2000s, businesses began realizing the wealth of insight into user behaviour web analytics applications could provide. The advent of free tools like Google Analytics and Piwik made analytics available to everyone.  This solidified the strategic role of analytics as more businesses adopted data-driven decision-making processes. Analyzing the effectiveness of digital initiatives now seems to be a no-brainer to stay in the game.

The single most important thing I can share with you:  it’s imperative to understand how the web data collection and measurement process works before
jumping into data reporting and analysis.


Today, web analytics is commonly done in two ways:

Log File Analysis: Web servers write details about user activity to a log file.  Those log file entries are then processed into database tables.  The reporting tool generates reports from these database tables. Although web analytics started with analyzing server log files, the measurement technique has become more sophisticated as analytics itself has matured.  Log file analysis remains a popular and accepted way of capturing data from websites.

JavaScript Coding: The most accepted method of data collection, however, is done through JavaScript coding. JavaScript code, which is placed in the HTML text of web pages, captures and stores data about a user’s activity.  This data is then sent to a collection server which stores it in database tables.  The reporting tool generates reports from these database tables.

So which method is better?


The preferred web analytic industry method is the JavaScript solution for a few reasons. First, server logs tend to report inflated page views and visit counts due to search engine indexing bots and spiders.  Not only do they depict false user traffic, but you get dinged and charged for page views every time your data is processed. Second, while log files capture server activity (page loads, etc.), they are unable to capture user interactions like button clicks, page scrolling, and form fallout, and they do not capture

actions within rich media or Flash content.  This limits the breadth and depth of analysis that’s possible.

How do LogFile and JavaScript analysis differ in unique visitor tracking?

Another limitation of log file analysis is unique visitor tracking. Since log files record web server activity, not web browser activity, they are not able to identify unique visitors –the same person (browser) requesting multiple pages from the server.  It’s actually this limitation that gave rise to JavaScript data collection methodology.

JavaScript data collection uses browser cookies to identify the same browser (person) across multiple page views and multiple visits.  (What’s a cookie? A cookie is a small text file that is assigned to a user’s browser and is stored on a user’s computer. That text file contains a name or ID number which identifies the user.  That identifier is transmitted to the data collection server every time they return to your site and as they navigate through it.)


Without a cookie to identify them, a person can visit your website 1000 X and you would never know the exact number of visitors you encountered. Here’s a simple formula to remember, no cookie = no tracking = crappy web analytics.

Not only that, modern Internet Service Providers (ISPs) have their own servers to cache web pages.  Caching means the ISP stores a temporary copy of your web page on their server.  They do this because it’s more efficient.  ISPs’ caching servers are usually faster to respond than most web servers.  The ISP’s server is probably located closer to the user’s computer than your web server, reducing slow speed caused by long-distance geography. Most important though, if two users request the same page, or the same user navigates back to it, they don’t have to load the page again from your web server.  When that happens, your log files have no record of it since there was no request to your web server.  You miss out on capturing that person’s activity history.

So, for those institutions using old school “Webtrends http: log file analysis,” it is strongly advised that you migrate to a JavaScript solution!

But that’s it for this blog folks; tune-in soon for a deep look at how the JavaScript collection process works!


Leave a Reply

Your email address will not be published. Required fields are marked *

Explore Posts By Category



Want to know more?

Contact us