Posts

Showing posts with the label hadoop

FREE Pivotal Training Courses

Image
Introduction to Pivotal CF Overview:   This course provides students with the concepts and hands-on experience needed to work with and deploy applications on Cloud Foundry. Students will gain familiarity in CF concepts (applications, buildpacks, manifests, organizations, spaces, users, roles, domains, routes, services), how to push applications to CF (various languages), services, user provided services, manifests, YAML, environment variables, autoconfiguration, logging and loggregator. Prerequisites:   Basic IT/computer science knowledge  High-level familiarity with web application architecture, from either a developer or administrator’s perspective. Intended Audience:   Developers Administrators Introduction to Pivotal HD   Overview:   This course is designed to provide an introduction to students new to Pivotal HD. As a one day course, students will gain a high level overview of features found in the Pivotal HD Hadoop distributi...

Sessions by Dr. Milind Bhandarkar, Chief Scientist at Pivotal

Image
At the last EMC Academic Conference in India, Dr. Milind Bhandarkar delivered 2 talks.  Here is a short profile of the speaker and links to his sessions. Please feel free to share it – ------------------------------------------------------------------------------------------------------------ Dr. Milind Bhandarkar was the founding member of the team at Yahoo, that took Apache Hadoop from 20-node prototype to datacenter-scale production system, and has been contributing and working with Hadoop since version 0.1.0. He started the Yahoo! Grid solutions team focused on training, consulting, and supporting hundreds of new migrants to Hadoop. Parallel programming languages and paradigms has been his area of focus for over 20 years. He worked at the Center for Development of Advanced Computing (C-DAC), National Center for Supercomputing Applications (NCSA), Center for Simulation of Advanced Rockets, Siebel Systems, Pathscale Inc. (acquired by QLogic), Yahoo! and Linkedin. C...

Hadoop Overview & Architecture

Hadoop Overview & Architecture from EMC Academic Alliance

Complete List of Big Ideas Videos from EMC Corp

Image
Big Idea videos from EMC are a great example of how to  Learning can be FUN . These  hand-drawn animations are created and narrated by EMC's Patricia Florissi, VP and Global Sales CTO. An Exhaustive list of these popular videos are below :   How Big is Big Data?  Big Ideas: Simplifying Cluster Architectures  Big Ideas: Demystifying Hadoop  Big Ideas: Simplifying High Performance Computing Architectures Big Ideas: Demystifying Fast Data  Big Ideas: Why Big Data Matters  Big Ideas: Simplifying Big Data Loading Big Ideas: Demystifying the Convergence of Big Data Architectures Big Ideas: When Data Needs Continuous Availability Big Ideas; Big Tech: Continuous Operations for VMware with EMC VPLEX Big Ideas; Big Tech: Continuous Operations for Oracle RAC with EMC VPLEX FLASH In A Flash Big Tech...

Three key predictions in Business Intelligence #GartnerBI

Image
Gartner has published a study on the need to scale up Business Intelligence and Analytics to Support Explosive Growth in Data Sources via areport titled  " Predicts 2013: Business Intelligence and Analytics Need to Scale Up to Support Explosive Growth in Data Sources ."  Some of the key finding are -  By 2015, 65 percent of packaged analytic applications with advanced analytics will come embedded with Hadoop. By 2016, 70 percent of leading BI vendors will have incorporated natural-language and spoken-word capabilities. By 2015, more than 30 percent of analytics projects will deliver insights based on structured and unstructured data. Source: gartner.com via EMC Academic Alliance on Pinterest

What is Greenplum HD ?

Image
Greenplum HD is enterprise-ready Apache Hadoop from EMC that allows users to write distributed processing applications for large data sets across a cluster of commodity servers using a simple programming model. This framework automatically parallelizes Map Reduce jobs to handle data at scale, thereby eliminating the need for developers to write  scalable and parallel algorithms.  Greenplum HD is an open source Apache stack and includes the following components: Hadoop Distributed File System (HDFS): File system that distributes files  across the cluster.  MapReduce: Framework for writing scalable data applications.  Pig: Procedural language that abstracts lower level MapReduce.  Hive: Data warehouse infrastructure built on top of Hadoop.  HBase: Database for random, real time read/write access.  Mahout: Scalable machine learning and data mining library.  ZooKeeper: Hadoop centralized servi...

What is "The Human Face of Big Data" Project

Image
“ The Human Face of Big Data ,” the latest groundbreaking, globally crowdsourced initiative from Rick Smolan, the creator of the “Day in the Life” series. The project, made possible through primary sponsorship from EMC, is based on the premise that the real-time visualization of data collected by satellites, and by billions of sensors, RFID tags, and GPS-enabled cameras and smartphones around the world, is enabling humanity to sense, measure, understand and affect aspects of our existence in ways our ancestors could never have imagined in their wildest dreams.  The multifaceted project kicks off on September 25 with an eight-day “Measure Our World” event inviting people around the world to share and compare their lives in real time through an innovative smartphone app. The project also includes “Mission Control” events in New York, Singapore and London; “Data Detectives,” a global student initiative being conducted in conjunction with the TED organization; a stunning large-...