Vertica Blog

Vertica Blog

ETL

Vertica and Attunity Replicate Change Data Capture

Attunity Replicate is a software solution that loads data from multiple sources and accelerates data integration for analytics, replication, streaming, and synchronizing, on-premises and in the cloud. Attunity Replicate and Vertica integration exists for Vertica as a Target. This integration enables moving large volumes of data from multiple sources into the Vertica database. It leverages...
Larry Bird Monument - Photo by Eric Kilby WC

Practice Makes Perfect at BDC 2020

Larry Bird, Boston Celtics legend and NBA Hall of Famer, shot 500 free throws every morning before his first class in high school. But, he didn’t stop there. Larry was known to shoot 100 free throws after every NBA game. The result? The sharp-shooting forward from French Lick, Indiana led the NBA in free throw...
partner fist bump

Vertica Partners Are the Key to Your Success at BDC 2020

There are a lot of reasons to be excited about the Vertica Big Data Conference 2020. One of my favorite parts of the event will be introducing our partner ecosystem to our customers! As Vertica’s partner marketing manager, one of my guiding principles is that innovation doesn’t happen in a silo.
Helpful Tips message on post-it note

DIRECT Is Now the Default Load Type

Write Optimized Store (WOS) is a memory-resident data structure for short-term data storage while Read Optimized Store (ROS) is a highly optimized, read-oriented, disk storage structure, organized by projection. Prior to Vertica 9.3, by default, Vertica initially loads data into WOS. For databases created in version 9.3 and later, Vertica now uses a default load...

Fast Data Loading with Vertica

Curtis Bennett authored this post. Vertica is well known for its blinding query performance at big data scale, but it can also insert data at very high rates of speed. It can even load data non-stop while being queried, thus enabling real-time analysis of data. Basic Loading Methods There are many ways of loading data...

What are your Data Loading Preferences?

We’re back with our newest product management survey for this summer! This time we’re asking about how you load your data – everything from the ETL tools you use to how you manage your information. The answers you provide will help us fit Vertica into your infrastructure in a way that is ideal for your...

Rebalance Taking a Long Time

After you add a node to your Vertica cluster or remove a node from your cluster, Vertica rebalances the data across all the nodes. If rebalancing is taking a long time, review these steps to find out the probable cause. Pre-Requisites To ensure a successful rebalance of your cluster, before you start the rebalance, take...

What Should I do if my Node Recovery is Slow?

If you are running Vertica 7.2.x or later, perform recovery by table. For details, see in the Vertica documentation. If you are running a Vertica version prior to 7.1.x, stop the ETL jobs and restart node recovery. Step Task Results 1 Monitor progress of recovery: If is_running = f, recovery completed. Repeat this statement to...

Sizing Your Vertica Cluster for an Eon Mode Database

This blog post was authored by Shrirang Kamat. Vertica in Eon Mode is a new architecture that separates compute and storage, allowing users to take advantage of cloud economics that enable rapid scaling and shrinking of clusters in response to a variable workload. Eon Mode decouples the cluster size from the data volume and lets...

Handling Duplicate Records in Input Data Streams

This blog post was authored by Ravi Gupta. We have often found that sources or operational systems that provide data for further analysis have duplicate records and these are sent to a downstream application or EDW for processing. This post shows a few scenarios of how to handle these duplicate records using various SQL options,...
Commercial passenger plane with Vertica painted on the tail

Blog Post Series: Using Vertica to Track Commercial Aircraft in near Real-Time – Part 6

Part 6: Extract, Transform and Load ADS-B messages into Kafka I have discussed in previous blog posts the continuous stream of messages from aircraft transponders, captured and decoded using the DUMP1090 application, which we are planning on feeding into a series of Kafka topics, prior to loading into their corresponding tables in a Vertica database....

Introducing the Parallel Streaming Transformation Loader (PSTL) Solution

This blog post was authored by Soniya Shah. At Vertica, we understand how important it is that our customers can make decisions in near real time. Being able to do this not only requires the massive parallel processing that Vertica offers, but the ability to transform and ingest your data into Vertica as quickly as...