Atomic Media text

Atomic Media

How to archive your Universal Analytics historical data

How to archive your Universal Analytics historical data

Another Google Analytics 4 migration project deadline is fast approaching, and this deadline is hard set. On July 1, Google will delete all historical data from Universal Analytics properties. This deadline also affects Analytics 360 customers.

With little more than a month until the deadline, if you have not done so by now, your organization needs to prioritize archiving your historical data. There are three main phases I recommend for approaching this project.

Phase 1: Make a plan

Before archiving data, it’s important to decide:

What specific data is important to you? 

How many years of data do you want to keep?

At what cadence do you review data?

Get the daily newsletter search marketers rely on.


See terms.


Phase 2: Choose an archiving method

There are four main options available for archiving your Universal Analytics data. Each has its own pros and cons, so choose a method based on your team’s resources and skills.

Option 1: Manual file downloads

While this is the easiest process to understand, it is also time consuming.

Following your plan for years, cadence and data points, you’ll need to go into each report in the Google Universal Analytics interface, set the date, dimension and metric settings as needed.

Also, remember to change the number of rows from the default of 10 to the maximum of 5,000 rows to ensure you capture as much data as possible.

Click the export button and export data to a Google Sheet, Excel or CSV. Repeat this process until you have downloaded all of the data identified in your archive plan.

Option 2: Download data to Google Sheets using the Google Analytics add-on (best option for tech novices)

This option is fairly simple for most users to perform. Create a new Google Sheet and add the Google Analytics spreadsheetadd-on.

The add-on essentially uses the Google Analytics API to download data to Google Sheets but doesn’t require API programming knowledge to operate. Google has compiled a basic overview of this approach in this help document.

The first time you use the add-on, you’ll build a report using the add-on’s interface. But after the first report has been run, you can also simply update the Report Configuration tab and create additional reports directly in columns of that sheet.

You can also conveniently use formulas in the Report Configuration sheet. Use the Dimensions and Metrics Explorer to find the proper API code to enter into each field. 

One drawback of the Google Sheets method is that you may encounter sampling if you pull too much data at once (e.g., your entire 20-year dataset for sessions) or your report is too detailed (too many dimensions pulled together for a high level of granularity). 

When you run a report, you’ll see the sampling level on the report’s data tab in cell B6. If your report contains sampled data, you may want to consider reducing the amount of data in this particular pull, for example, you might split the pull into two time frames.

However, if you just can’t avoid sampling, check the data sample percentage on the report. Then, on the Report Configuration tab, unhide rows 14-17 and the sampling size on row 15 to this level so that your data remains consistent.

Tip: The add-on defaults to 1,000 lines of data in a report. Simply delete the 1,000 under the line labeled “Limit” (typically row 11).

Another drawback of the Google Sheets option is that each file is limited to 10,000,000 cells. Typically, each sheet starts out with 26 columns (A to Z) and 1,000 default rows (or 26,000 cells).

If your downloaded data exceeds the 10,000,000 cell limitation (which can very likely happen), then you may need to have multiple Google Sheets to download all of the data.

Option 3: Download data using the Google Analytics API

If you have web development resources that can work on the archiving project, they can pull the data detailed in your plan using the Google Analytics API directly. 

This works similarly to the aforementioned Google Sheets add-on option, but it’s a more manual process in programming the API calls.

To learn about how to use the API for this project, visit Google’s archiving information page and review the second bullet, which details several resources and considerations for using the API for this data export project.

Option 4: Download data to BigQuery (best option overall)

The main benefit of archiving your Universal Analytics data to BigQuery is that BigQuery is a data warehouse that allows you to ask questions of the data set through SQL queries to get your data very quickly. This is especially useful in accessing this data for reporting later. 

Analytics 360 users

If you are an Analytics 360 user, Google provides a native export to BigQuery. I recommend this method. See instructions from Google.

Everyone else

If you’re not an Analytics 360 user, then you’ll need to approach the BigQuery backup differently because Google does not provide innate BigQuery backup options in Universal Analytics for non-360 users. 

Here are the steps you’ll want to follow:

Because you only need to transfer the Universal Analytics data one time, you can also change the schedule on the transfer to On demand and then run the transfer now. 

Phase 3: Ensure you’ve captured it all

Before you consider the project complete, be sure to double-check your archived data to ensure you’ve captured everything you planned to archive.

On July 1, you will no longer be able to access Universal Analytics data, either by API or through the interface. 

Courtesy of Search Engine Land: News & Info About SEO, PPC, SEM, Search Engines & Search Marketing

Category seo news | Tags:

Social Networks : Technorati, Stumble it!, Digg, de.licio.us, Yahoo, reddit, Blogmarks, Google, Magnolia.

You can follow any responses to this entry through the RSS 2.0 feed.

No Responses to
“How to archive your Universal Analytics historical data”





XHTML: You can use these tags: <a href="" title=""> <abbr title=""> <acronym title=""> <b> <blockquote cite=""> <cite> <code> <del datetime=""> <em> <i> <q cite=""> <strike> <strong>

By submitting a comment here you grant Atomic Media a perpetual license to reproduce your words and name/web site in attribution. Inappropriate comments will be removed at admin's discretion.