Vertica Blog
Sarah Lemaire smiling

Sarah Lemaire

Manager, Vertica Documentation

What are Vertica Engineers Reading?

Ever wonder what Vertica engineers do when they aren?t busy building and perfecting an awesome product?

Restoring Objects from a Full Backup

Before Vertica version 7.2, you had two options: perform full backups and restore full backups (remove everything from the storage unit) or perform object-level backups and restore those individually (maintain separate storage units). But as of Vertica 7.2, you can have the best of both worlds. You can perform full backups and restore individual objects...
Modern Database Analytics

Understanding ROS and WOS: A Hybrid Data Storage Model

Vertica is unique in many ways, one of which can be seen in its data storage model. In this blog, we?ll go over each storage component and show you how to take advantage of Vertica?s storage options to optimize your database for writing and reading.
Database Server Room

Using Big Data Analytical Tools for Predicting Potential Global Distribution of an Amphibian Pathogen

During the summer of 2015, I participated in an internship program with Vertica. Most interns assisted in software development, but my primary goal was to use Vertica, Vertica Place, and HP Distributed R to address an ecological problem

Kerberos Authentication for Hadoop Integration: Keeping your data safe

Like all Vertica releases, 7.1.2 includes a bunch of new features. One specific feature can help make your HDFS data safer: Kerberos integration for Hadoop. This blog will give you an overview of using Kerberos authentication with Vertica for SQL on Hadoop.

Vertica and Qlik Team Up to Hack for a Cause

On November 17 2014, Vertica took part in the annual Qlik Hackathon in Orlando, Florida. The hackathon, which took place the day before the Qlik World Conference, brought together engineers, developers, and Qlik experts from all over the world. The challenge was simple: leverage features from Qlik and its partners to create an application to...

Best Practices for Refreshing Large Projections

More than once I have worked with customers who need to update a superprojection or create a new projection for a large fact table. It seems like a simple and easy process: just create a projection and perform a refresh. However, refreshing projections for large fact tables can produce unwanted complications. In this blog, we?ll...

What is a Lock Anyway?

We talk a lot about database security and how you can protect your sensitive data from outside threats. But what about internal, unintentional data corruption? What if the data you are trying to analyze or manipulate is simultaneously being manipulated by another transaction? A scenario such as this could lead to data loss and inconsistency....

Time Series Analytics: The Hunt for “Missing Link” in Data

Time series analytics is a little-known, but very powerful Vertica tool. In Vertica, the TIMESERIES clause and time series aggregate functions normalize data into time slices. Then they interpolate missing values that fill in the gaps.Using time series analytics is useful when you want to analyze discrete data collected over time, such as stock market...
Three 3D arrows, different colors pointing in different directions

Always Have a Backup Plan

Why should you always have a backup plan? Because something, inevitably, goes wrong. Whether you experience something as simple as a deleted record, or as serious as a full system deletion, you need a backup.

For Your Eyes Only: Using Vertica Access Policies

Everyone knows there?s virtually no limit to the amount of data you can store, manage, and query in Vertica. But you may not be aware of how easy it is to control who sees what in your database. In this blog, Alexey Kireyev, the lead security developer for Vertica, explains how you can use the...
Database Server Room

Vertica Best Practices: Identifying Queries for DBD

The Vertica Database Designer is a tool that analyzes a logical schema definition, sample queries, and sample data, and creates a physical schema (projections) in the form of a SQL script that you deploy automatically or manually. The result is an optimized database with optimal query performance and data compression.