Talend, a provider of cloud data integration and data integrity, is releasing its Summer '19 version of Talend Data Fabric, accelerating the development of all types of integration environments from simple ingestion tasks to the most comprehensive integration scenarios.
Posted July 17, 2019
Eight Trends in Data Analytics
Posted July 17, 2019
Increasingly, people across the spectrum of organized human activity—from business to government—are recognizing the importance of better managing and governing their data assets. "Data is becoming an increasingly critical foundation of the economy and of our lives," said Kevin Lewis, director of strategic offer management for Teradata. "The more this happens, the more important regulation will be, not just for privacy, but also for data quality."
Posted July 17, 2019
IGEL, a provider of the next-gen edge OS for cloud workspaces, and CloudJumper are forming a partnership to simplify the deployment and management of cloud workspaces, including WVD workspaces delivered on Microsoft Azure.
Posted July 16, 2019
Logi Analytics Snaps Up Zoomdata, Extending Leadership in Embedded Analytics
Posted June 18, 2019
Data visualization is often described as part art and part science, and it's true: Visual design and information representation are separate yet highly interrelated concepts that, together, create meaning visually from data. Applied in data visualization, these elements leverage our brain's cognitive functions to help us better see and understand information, and to interact with, learn from, and reach new insights in our data.
Posted May 16, 2019
Apache Airflow is turning heads these days. It integrates with many different systems and it is quickly becoming as full-featured as anything that has been around for workflow management over the last 30 years. This is predominantly attributable to the hundreds of operators for tasks such as executing Bash scripts, executing Hadoop jobs, and querying data sources with SQL.
Posted May 16, 2019
Security operation centers (SOCs) are grappling with the demands of a growing threat landscape.
Posted May 16, 2019
Becoming future-proof sounds similar to something you would need a fortune teller for. If anyone can teach me how to become a fortune teller, then I would like to sign up for that training class! But I don't believe that you have to be a fortune teller to be future-proof. Rather, it is more about acknowledging some hard facts about the pitfalls in your organization, and then trying to avoid stepping in them.
Posted May 16, 2019
It is still early for the use of cognitive technologies and AI, but many organizations are exploring the potential it holds across a range of industries. Retail, banking, healthcare, and manufacturing are industries leading the charge to leverage AI today. At the same time, concerns persist.
Posted May 16, 2019
Data architectures are becoming more complex and changing more frequently, requiring that a new operational mindset be applied to data management. In particular, automating the building and maintenance of data pipelines is needed, as is instrumenting and continuously monitoring pipeline performance to ensure reliability and quality for data consumers. We call this practice "DataOps."
Posted May 16, 2019
Keep your eye on the ball! Good eye! Way to watch! Baseball season is in full swing (pun maybe intended) and, with three sons, I have begun to hear these phrases in my sleep. Enough so that it sent me down a rabbit hole of curiosity. I began looking for the origin of the phrase "Keep your eye on the ball." Turns out, there is an interesting parallel with data governance.
Posted May 16, 2019
Ask 10 people "What is high availability?" and you might get 15 different answers. For the purposes here, HA will be defined as ensuring SQL Server is operating when and as needed. "When" takes into account the percentage of time SQL Server is up and running, while "as" takes into account proper operation with no data loss and/or corruption.
Posted May 09, 2019
Next Stop, Silicon Valley Africa
Posted May 09, 2019
Accenture's Bhaskar Ghosh recently reflected on how systems have and will change and the forces driving this transformation. "Boundaries between applications and infrastructure are blurring to the point they're nearly indistinguishable," noted Ghosh.
Posted May 09, 2019
Kubeflow is a workflow tool which prides itself on making machine learning workflows simple to build, scalable, and portable. It provides graphical end-user tools to set up and define the steps in a pipeline. Most importantly, as data scientists build out their use cases, they add more and more steps and, when using Kubeflow, they end up with a documented, repeatable process.
Posted April 24, 2019
Cloud computing—and everything that goes with it—is dramatically changing the roles and aspirations of database administrators. No longer do DBAs need to be chained to their databases, wrestling with managing updates, applying security patches, and dealing with capacity issues. Moving to a cloud data environment is steadily shifting DBAs' roles from hands-on database overseers to value-drivers for their businesses—and enabling a range of career advancement opportunities not seen since the dawn of relational databases.
Posted April 11, 2019
Your mileage may vary, but there is an allegorical household phenomenon that provides some lessons in understanding the data lake—the coffee can that often ends up in the garage or tool space of a home to collect all the stray, old-project hardware someone in the household cannot bear to part with.
Posted April 02, 2019
What happened to the 500 million data points from the Starwood data breach?
Posted April 02, 2019
How Containers and Microservices Enable Cloud's Next Frontier: Q&A with IBM's Jim Comfort
Posted April 02, 2019
It's easy to see when the emperor has no clothes, just as it's easy to spot a truly bad technology. What's much harder is spotting an overhyped technology—one that has great promise that hasn't been fulfilled yet, or one that is great for a given purpose, but positioned as the cure for world hunger.
Posted March 27, 2019
Working Backward (Or: What IoT Can Learn From Steve Jobs)
Posted March 08, 2019
Good data visualization should be fast, informative, and—above all—valuable. This makes data viz a critical tool in the modern analyst toolkit. Here are 3 simple questions to ensure your data visualization passes the eye candy test.
Posted March 06, 2019
Enterprise application teams are facing pressure to release applications more quickly, but most enterprises still have a manual process for reviewing, validating, and deploying database changes. This creates a bottleneck for business innovation and improving CX.
Posted March 06, 2019
Bias—whether consciously or unconsciously introduced—can ruin any project and lead to extremely damaging conclusions if inaccurate or incomplete data is fed in. We've already seen multiple examples of engineers being unable to prevent AI networks from accidentally becoming racist or sexist, and it's not hard to see how bias can lead to many worrying outcomes.
Posted March 06, 2019
Improving Db2 Performance: Q&A with Craig S. Mullins
Posted February 26, 2019
While blockchain is believed to be a solution to many business problems, this is not necessarily the case. But, it is igniting new discussions that can help enterprises find the right solutions to some of their biggest challenges. In all the hype, people tend to focus on technical discussions and fail to realize the value in just how much this technology is transforming the way we think within the enterprise.
Posted February 19, 2019
For many companies, designing and implementing a data platform for analytics is a critical task. According to Dremio, a VC-backed firm founded in 2015, by combining capabilities and technologies into a solution that enables access, transformation, security, and governance, data as a service represents a new approach to vexing analytics challenges, delivering data at scale with high performance. Recently, Kelly Stirman, vice president strategy at Dremio, discussed how using open source projects, open standards, and cloud services, companies can deliver data as a service to their data consumers across critical lines of business.
Posted February 12, 2019
Veritas Technologies, a provider of enterprise data protection and software-defined storage, and its platforms Veritas NetBackup and Veritas Backup Exec have attained Amazon Web Services (AWS) Storage Competency status, reaffirming Veritas as an AWS Partner Network (APN) Advanced Technology Partner offering solutions validated by the AWS Storage Competency.
Posted February 12, 2019
Logi Analytics has acquired Jinfonet Software, maker of JReport. The acquisition will consolidate two leading embedded analytics vendors under the Logi brand.
Posted February 12, 2019
Attunity, a provider of data integration and big data management software solutions, has announced two new solutions: Attunity for Data Lakes on Microsoft Azure, designed to automate streaming data pipelines, and Attunity Compose for Microsoft Azure SQL Data Warehouse, designed to enable data warehouse automation for Azure SQL Data Warehouse.
Posted February 12, 2019
MapR Technologies, provider of a data platform for AI and analytics, has announced support for Apache Drill 1.15. The new release offers new enhancements to conduct queries on complex nested data structures, including files, MapR JSON database tables, and cloud data sources specifically for S3 (Amazon Simple Storage Service?).
Posted January 30, 2019
AI: Data Quality's New Frontier
Posted January 09, 2019
The State of Cloud Technologies: The Past Controls the Future
Posted January 09, 2019
The Importance of Data for Applications and AI
Posted January 09, 2019
Steering Your Data Ship with Data Governance
Posted January 09, 2019
Don't Get Washed Out by the Overflowing Data Lake: 5 Key Considerations
Posted December 13, 2018
Regulatory Compliance Demands Sustainable Information Governance
Posted December 12, 2018
3 Ways Machine Learning Is Transforming Data Analytics
Posted December 12, 2018
The State of Software Licensing: The Monetization Playbook
Posted December 11, 2018
Fast and Furious Technology: From HFT to IoT
Posted December 11, 2018
The newest edition of the TOP500 list of the world's fastest supercomputers puts five U.S. Department of Energy (DOE) supercomputers in the top 10 positions, with the first two captured by Summit at Oak Ridge National Laboratory (ORNL) and Sierra at Lawrence Livermore National Laboratory (LLNL).
Posted November 12, 2018
Arun Murthy—CPO and co-founder of Hortonworks—who will also serve as CPO at Cloudera after the merger is complete—recently shared his thoughts on what's ahead for 2019. According to Murthy, data at the edge, AI, IoT, open source, and cloud will all factor in strongly in organizations' plans for analytics and governance.
Posted November 12, 2018
The data lake has become accepted as an important component of a modern data architecture, enabling a wider variety of data to be accessed by users. Yet, challenges persist. Recently, John O'Brien, CEO and principal advisor, Radiant Advisors, talked about the cultural transformation underway as companies increasingly realize the power of data, and the tools and technologies helping them to expand what is possible.
Posted October 24, 2018
In my last article, I discussed how the Internet of Things market is showing early signs of maturity, but that many projects still can stumble. I identified seven "habits" that successful projects have in common, which, when used together, are powerful enough to set your IoT project on the right path.
Posted September 27, 2018
Volumes of data surround us. The internet, advertising, social media, connected cars and smart homes have driven an exponential increase in information, primarily unstructured data. But, this data is useless unless we can also comprehend, analyze and make operational, tactical and strategic (yet democratized) decisions based on the information. This realization has pushed organizations to rethink business strategies and outcomes as related to big data and digital transformation. It's also prompted an investment in technologies that enable data mining, as well as prescriptive, descriptive and predictive analytical solutions.
Posted September 27, 2018
Thanks to the dramatic uptick in GPU capabilities, gone are the days when data scientists created and ran models in a one-off manual process. This is good news because the one-off model was typically not optimized to get the best results. The bad news is that with an increase in the total number of models created—including iterations over time—the amount of data used as inputs and generated by the models quickly spirals out of control. The additional bad news is that there are a variety of complexities associated with model, data, job, and workflow management.
Posted September 26, 2018
This past summer, I had the opportunity to stay in a wide variety of hospitality establishments for both personal and professional travel, which, for me, has generated another fun way to look at data governance that I am excited to share. Because, let's face it, without good analogies, data governance on its own can be, well, kind of dry. Welcome to the Hotel Data Governance. Such a lovely place. (The tune is already in your head isn't it? You're welcome.)
Posted September 26, 2018
Many organizations nowadays are struggling with finding the appropriate data stores for their data, making it important to understand the differences and similarities between data warehouses, data marts, ODSs, and data lakes. All these data structures clearly serve different purposes and user profiles, and it is necessary to be aware of their differences in order to make the right investment decisions.
Posted September 26, 2018
Both Oracle and SQL Server have very well-established communities. While they are different, they are also similar in many ways. All DBAs worry about the performance and security of the data and the database. Out of necessity, Oracle DBAs have become more specialized. Will this happen to SQL Server DBAs now that the database is offered on Linux?
Posted September 25, 2018