Get Instant Access
to This Blueprint

Applications icon

Choose the Right Tools for Big Data Development

Leverage Hadoop as your pilot project to gain organizational buy-in and build institutional learning.

  • A relational database management system works great under many scenarios, but it has its limitations:
    • Volume issues typically arise when there is a need to index large databases.
    • In a multi-source environment, data collisions occur and resolving them can be expensive and time consuming.
    • Velocity problems arise when large amounts of read/write transactions occur that are expensive to compute.

Our Advice

Critical Insight

  • Begin your big data implementation with a baseline Hadoop pilot. This pilot will help build your knowledge of big data, how the Hadoop framework satisfies your use cases, and how it operates in your system. Each component in this baseline stack is well understood in the industry and documentation is readily available.

Impact and Result

  • Provide a step-by-step starting point to begin the rollout of big data development based on your business and technical requirements.
  • Highlight the challenges, impacts, potential, and mitigations in big data development.
  • Identify the key metrics, benchmarks, and instrumentation points to measure the success of your big data rollout.

Choose the Right Tools for Big Data Development Research & Tools

1. Assess fit and readiness for big data

Minimize the process and technology impacts of introducing big data to the organization.

2. Build the project team

Identify the roles and responsibilities of the big data project team.

3. Roll out the Hadoop pilot

Create a Hadoop stack based on business requirements and the data that needs to be mined and analyzed.

4. Roll out Hadoop in the organization

Customize the Hadoop pilot for fit in other areas of the organization based on instrumentation and pilot experiences.

Leverage Hadoop as your pilot project to gain organizational buy-in and build institutional learning.

About Info-Tech

Info-Tech Research Group is the world’s fastest-growing information technology research and advisory company, proudly serving over 30,000 IT professionals.

We produce unbiased and highly relevant research to help CIOs and IT leaders make strategic, timely, and well-informed decisions. We partner closely with IT teams to provide everything they need, from actionable tools to analyst guidance, ensuring they deliver measurable results for their organizations.

What Is a Blueprint?

A blueprint is designed to be a roadmap, containing a methodology and the tools and templates you need to solve your IT problems.

Each blueprint can be accompanied by a Guided Implementation that provides you access to our world-class analysts to help you get through the project.

Need Extra Help?
Speak With An Analyst

Get the help you need in this 1-phase advisory process. You'll receive 3 touchpoints with our researchers, all included in your membership.

  • Call 1: Assess fit and readiness for big data

    Get off to a productive start: Assess your data analytics stacks to determine your readiness for big data. Info-Tech analysts will help you identify your gaps and create a list of tasks to fill these gaps.

  • Call 2: Prepare and roll out the Hadoop pilot

    Build your Hadoop pilot stack: Review the roles and responsibilities for your Hadoop pilot, document your requirements, choose an installation approach, design and build your MapReduce program, select your analytics toolset, and test and validate your Hadoop data flow. Info-Tech analysts will discuss the fit of your Hadoop stack to your business requirements and assist in planning for the stack implementation.

  • Call 3: Roll out Hadoop in the organization

    Monitor and optimize your Hadoop stack for deployment: Identify your instrumentation points and metrics, tweak your Hadoop stack based on your instrumentation to fit other areas of your organization, and apply lessons learned to other development projects. Info-Tech analysts will discuss the success of your big data rollout and help you optimize your big data stack.


Andrew Kum-Seun

Altaz Valani


Individuals who conducted expert interviews with us for this project:

  • Martin Parrest, Foxnet Solutions
  • Mehdi Bahrami, University of California, Merced
  • Michael Hausenblas, MapR Technologies
  • Michael Davison, Davis+Henderson

Vendors who conducted expert interviews with us for this project:

  • Pentaho
  • Informatica

Search Code: 73868
Last Revised: May 19, 2014

Visit our Exponential IT Research Center
Over 100 analysts waiting to take your call right now: 1-519-432-3550 x2019