|Course or Certification Name||Category||Location||Mode of learning|
|The Ultimate Hands-On Hadoop - Tame your Big Data!||Big Data||Online self study|
|Advanced Operating Systems||Relational Databases||Online self study|
|Introduction to Cloud Foundry and Cloud Native Software Architecture||Cloud Computing||Online self study|
|Cloud Computing Concepts Part 2||Cloud Computing||Online self study|
|Cloud Computing Concepts Part 1||Cloud Computing||Online self study|
|CISSP (Online Classroom - Flexi Pass)||cissp||Noida , Delhi , Gurgaon , Chandigarh , Bangalore , Hyderabad , Chennai , Ernakulam||Online Classroom|
|Big Data Hadoop Architect Masters Program||Hadoop Administration||Noida , Delhi , Gurgaon , Chandigarh , Bangalore , Hyderabad , Chennai , Ernakulam||Online Classroom|
|AWS Developer: Building on AWS||Cloud Computing||Online self study|
|Cloud Computing Specialization||Cloud Computing||Online self study|
|NoSQL||NO SQL Databases||Online self study|
|Apache Spark and Scala (Online Classroom-Flexi Pass)||Big Data||Noida , Delhi , Gurgaon , Chandigarh , Bangalore , Hyderabad , Chennai , Ernakulam||Online Classroom|
|Big Data Expert||Big Data||Noida , Delhi , Gurgaon , Chandigarh , Bangalore , Hyderabad , Chennai , Ernakulam||Online Classroom|
|Managing Big Data in Clusters and Cloud Storage||Big Data||Online self study|
|SAP s/4hana conversion and sap system upgrade||SAP BI||Online self study|
|DevOps Plus Training||DevOps||Noida , Delhi , Gurgaon , Chandigarh , Bangalore , Hyderabad , Chennai , Ernakulam||Online Classroom|
The world of Hadoop and "Big Data" can be intimidating - hundreds of different technologies with cryptic names form the Hadoop ecosystem. With this Hadoop tutorial, you'll not only understand what those systems are and how they fit together - but you'll go hands-on and learn how to use them to solve real business problems! Learn and master the most popular big data technologies in this comprehensive course, taught by a former engineer and senior manager from Amazon and IMDb . We'll go way beyond Hadoop itself, and dive into all sorts of distributed systems you may need to integrate with. Install and work with a real Hadoop installation right on your desktop with Hortonworks (now part of Cloudera) and the Ambari UI Manage big data on a cluster with HDFS and MapReduce Write programs to analyze data on Hadoop with Pig and Spark Store and query your data with Sqoop , Hive , MySQL , HBase , Cassandra , MongoDB , Drill , Phoenix , and Presto Design real-world systems using the Hadoop ecosystem Learn how your cluster is managed with YARN , Mesos , Zookeeper , Oozie , Zeppelin , and Hue Handle streaming data in real time with Kafka , Flume , Spark Streaming , Flink , and Storm Understanding Hadoop is a highly valuable skill for anyone working at companies with large amounts of data. Almost every large company you might want to work at uses Hadoop in some way, including Amazon, Ebay, Facebook, Google, LinkedIn, IBM, Spotify, Twitter, and Yahoo! And it's not just technology companies that need Hadoop; even the New York Times uses Hadoop for processing images. This course is comprehensive, covering over 25 different technologies in over 14 hours of video lectures . It's filled with hands-on activities and exercises, so you get some real experience in using Hadoop - it's not just theory. You'll find a range of activities in this course for people at every level. If you're a project manager who just wants to learn the buzzwords, there are web UI's for many of the activities in the course that require no programming knowledge. If you're comfortable with command lines, we'll show you how to work with them too. And if you're a programmer, I'll challenge you with writing real scripts on a Hadoop system using Scala, Pig Latin, and Python . You'll walk away from this course with a real, deep understanding of Hadoop and its associated distributed systems, and you can apply Hadoop to real-world problems. Plus a valuable completion certificate is waiting for you at the end! Please note the focus on this course is on application development, not Hadoop administration. Although you will pick up some administration skills along the way. Knowing how to wrangle "big data" is an incredibly valuable skill for today's top tech employers. Don't be left behind - enroll now! "The Ultimate Hands-On Hadoop... was a crucial discovery for me. I supplemented your course with a bunch of literature and conferences until I managed to land an interview. I can proudly say that I landed a job as a Big Data Engineer around a year after I started your course. Thanks so much for all the great content you have generated and the crystal clear explanations. " - Aldo Serrano "I honestly wouldnâ€š be where I am now without this course. Frank makes the complex simple by helping you through the process every step of the way. Highly recommended and worth your time especially the Spark environment. This course helped me achieve a far greater understanding of the environment and its capabilities. Frank makes the complex simple by helping you through the process every step of the way. Highly recommended and worth your time especially the Spark environment." - Tyler Buck
In this course, we will see all the advances that have led to the state-of-the-art operating system that we know today, covering variety of platforms -- cell phones, multi-core, parallel systems, distributed systems, and cloud computing.
Does your team use Cloud Foundry to deploy applications? Or would you like to use Cloud Foundry, but haven't had time to learn the lingo? Then this course is just what you need! Cloud Foundry makes it simple for developers to deliver business value more quickly, without wasting time getting their app to the cloud -- it's already there. |This course is an introduction to Cloud Foundry, including distributions available to end users, an overview of the platform's components, and what it means to be Cloud Foundry certified. The course also includes technical instructions on how to use the command line interface, how applications are deployed, what services are within the context of the system and basic debugging practices.
Cloud computing systems today, whether open-source or used inside companies, are built using a common set of core techniques, algorithms, and design philosophies â€“ all centered around distributed systems. Learn about such fundamental distributed computing "concepts" for cloud computing. |Some of these concepts include: clouds, MapReduce, key-value/NoSQL stores, classical distributed algorithms, widely-used distributed algorithms, scalability, trending areas, and much, much more!
Cloud computing systems today, whether open-source or used inside companies, are built using a common set of core techniques, algorithms, and design philosophies â€“ all centered around distributed systems. Learn about such fundamental distributed computing "concepts" for cloud computing. | Some of these concepts include: clouds, MapReduce, key-value/NoSQL stores, classical distributed algorithms, widely-used distributed algorithms, scalability, trending areas, and much, much more!
Certified Information Systems Security Professional (CISSP) training facilitates the learning and knowledge in designing, building, and maintaining a secure business environment using globally approved Information Security standards. The aim of the course is to develop expertise in architecture, and design of security and earn the 30 CPEs required for the CISSP certification exam. IT Security professionals such as Security Consultants and Managers, IT Directors and Managers, and Security Auditors and Architects are best suited for this course enhancing their skills and knowledge on business continuity planning & disaster recovery planning vital for the smooth running of any organization
Big Data Hadoop Architect Program is a certification course that would help you build strong skill set in areas like Hadoop Development, Real time processing using Spark, and NoSQL database technology and transform you into a Hadoop Architect Expert. You would also be gain practical experience by implementing real life industry projects in the required Hadoop technologies.
In this course, part of the AWS Developer Series, you will develop and deploy applications on the AWS platform. Throughout the course, working in Python on Linux, you will develop a web application building upon your developer skills and using AWS services and tools. |Your AWS instructors will guide you through getting started with AWS, creating an account, and installing SDKs. Throughout the course, we will use hands-on exercises to build out a web application.
The Cloud Computing Specialization takes you on a tour through cloud computing systems. We start in in the middle layer with Cloud Computing Concepts covering core distributed systems concepts used inside clouds, move to the upper layer of Cloud Applications and finally to the lower layer of Cloud Networking. We conclude with a project that allows you to apply the skills you've learned throughout the courses.The first four courses in this Specialization form the lecture component of courses in our online Master of Computer Science Degree in Data Science. You can apply to the degree program either before or after you begin the Specialization.
Apache Spark is an open-source cluster-computing framework used for Big Data Processing. It combines SQL, streaming and complex analytics together seamlessly to handle a wide range of data processing scenarios. Scala is a general-purpose programming language which is supported by Apache Spark. This Apache Spark and Scala course is designed for candidates who want to advance their skills and expertise in Big Data Hadoop Ecosystem. Designed by experts in the industry, this course offers training on various topics like Spark Streaming, Spark SQL, Machine Learning Programming, GraphX Programming and Shell Scripting Spark. In addition to this, the candidates get to work on real life industry project. Upon completion of this course, successful candidates get experience certificate in Apache Spark and Scala.
Hadoop is an Apache project (i.e. an open source software) to store & process Big Data. Hadoop stores Big Data in a distributed & fault tolerant manner over commodity hardware. Afterwards, Hadoop tools are used to perform parallel data processing over HDFS (Hadoop Distributed File System) | As organisations have realized the benefits of Big Data Analytics, so there is a huge demand for Big Data & Hadoop professionals. Companies are looking for Big data & Hadoop experts with the knowledge of Hadoop Ecosystem and best practices about HDFS, MapReduce, Spark, HBase, Hive, Pig, Oozie, Sqoop & Flume | Edureka Hadoop Training is designed to make you a certified Big Data practitioner by providing you rich hands-on training on Hadoop Ecosystem. This Hadoop developer certification training is stepping stone to your Big Data journey and you will get the opportunity to work on various Big data projects.
In this course, you'll learn how to manage big datasets, how to load them into clusters and cloud storage, and how to apply structure to the data so that you can run queries on it using distributed SQL engines like Apache Hive and Apache Impala. Youâ€™ll learn how to choose the right data types, storage systems, and file formats based on which tools youâ€™ll use and what performance you need.
The "SAP Certified Technology Specialist - SAP S/4HANA Conversion and SAP System Upgrade" certification exam validates that the candidate possesses the required level of knowledge to perform a technical SAP S/4HANA conversion and an SAP system upgrade. This certification proves that the candidate has an overall understanding and the in?depth technical skills required. Armed with an understanding of SAP system administration, the candidate can implement this knowledge practically in SAP projects. The enablement for this exam is being provided by the course "ADM328 - SAP S/4HANA Conversion and SAP System Upgrade". It is highly recommended that you have passed the exam C_TADM55_74 or C_TADM55_75 or C_TADM55A_75 prior to taking this exam
The DevOps Plus course takes a comprehensive look at the discipline that covers all key concepts, methodologies, and tools, regardless of your understanding of the IT technologies and practices. Starting with a foundational introduction to DevOps, it covers the concepts of virtualization, its benefits, and the many virtualization products that play a significant role in both learning and implementing the DevOps culture. Youâ€™ll also learn about DevOps tools such as Vagrant, Version Control Systems, Docker, Containerization, and Configuration Management with Chef, SaltStack, Puppet, and Ansible. | This course focuses on both mid-level and advanced concepts, including open-source monitoring application Nagios, its plug-ins, and its use as a GUI. The CI/CD Pipeline Automation covers the Advanced DevOps concepts that are covered in detail along with Docker container clustering using Docker Swarm and Kubernetes.