ºÝºÝߣshows by User: PrashantGupta82 / http://www.slideshare.net/images/logo.gif ºÝºÝߣshows by User: PrashantGupta82 / Sun, 08 Oct 2017 04:40:51 GMT ºÝºÝߣShare feed for ºÝºÝߣshows by User: PrashantGupta82 Spark core /slideshow/spark-core-80569932/80569932 spark2-171008044051
Apache Spark is a fast and general engine for big data processing, with built-in modules for streaming, SQL, machine learning and graph processing.]]>

Apache Spark is a fast and general engine for big data processing, with built-in modules for streaming, SQL, machine learning and graph processing.]]>
Sun, 08 Oct 2017 04:40:51 GMT /slideshow/spark-core-80569932/80569932 PrashantGupta82@slideshare.net(PrashantGupta82) Spark core PrashantGupta82 Apache Spark is a fast and general engine for big data processing, with built-in modules for streaming, SQL, machine learning and graph processing. <img style="border:1px solid #C3E6D8;float:right;" alt="" src="https://cdn.slidesharecdn.com/ss_thumbnails/spark2-171008044051-thumbnail.jpg?width=120&amp;height=120&amp;fit=bounds" /><br> Apache Spark is a fast and general engine for big data processing, with built-in modules for streaming, SQL, machine learning and graph processing.
Spark core from Prashant Gupta
]]>
1284 2 https://cdn.slidesharecdn.com/ss_thumbnails/spark2-171008044051-thumbnail.jpg?width=120&height=120&fit=bounds presentation Black http://activitystrea.ms/schema/1.0/post http://activitystrea.ms/schema/1.0/posted 0
Spark Sql and DataFrame /slideshow/spark-sql-and-dataframe/80569923 sparksql-171008043944
Apache Spark is a fast and general engine for big data processing, with built-in modules for streaming, SQL, machine learning and graph processing.]]>

Apache Spark is a fast and general engine for big data processing, with built-in modules for streaming, SQL, machine learning and graph processing.]]>
Sun, 08 Oct 2017 04:39:44 GMT /slideshow/spark-sql-and-dataframe/80569923 PrashantGupta82@slideshare.net(PrashantGupta82) Spark Sql and DataFrame PrashantGupta82 Apache Spark is a fast and general engine for big data processing, with built-in modules for streaming, SQL, machine learning and graph processing. <img style="border:1px solid #C3E6D8;float:right;" alt="" src="https://cdn.slidesharecdn.com/ss_thumbnails/sparksql-171008043944-thumbnail.jpg?width=120&amp;height=120&amp;fit=bounds" /><br> Apache Spark is a fast and general engine for big data processing, with built-in modules for streaming, SQL, machine learning and graph processing.
Spark Sql and DataFrame from Prashant Gupta
]]>
324 5 https://cdn.slidesharecdn.com/ss_thumbnails/sparksql-171008043944-thumbnail.jpg?width=120&height=120&fit=bounds presentation Black http://activitystrea.ms/schema/1.0/post http://activitystrea.ms/schema/1.0/posted 0
Map Reduce /slideshow/map-reduce-79856653/79856653 mapreduce-170917102008
A MapReduce job usually splits the input data-set into independent chunks which are processed by the map tasks in a completely parallel manner. The framework sorts the outputs of the maps, which are then input to the reduce tasks. Typically both the input and the output of the job are stored in a file-system.]]>

A MapReduce job usually splits the input data-set into independent chunks which are processed by the map tasks in a completely parallel manner. The framework sorts the outputs of the maps, which are then input to the reduce tasks. Typically both the input and the output of the job are stored in a file-system.]]>
Sun, 17 Sep 2017 10:20:08 GMT /slideshow/map-reduce-79856653/79856653 PrashantGupta82@slideshare.net(PrashantGupta82) Map Reduce PrashantGupta82 A MapReduce job usually splits the input data-set into independent chunks which are processed by the map tasks in a completely parallel manner. The framework sorts the outputs of the maps, which are then input to the reduce tasks. Typically both the input and the output of the job are stored in a file-system. <img style="border:1px solid #C3E6D8;float:right;" alt="" src="https://cdn.slidesharecdn.com/ss_thumbnails/mapreduce-170917102008-thumbnail.jpg?width=120&amp;height=120&amp;fit=bounds" /><br> A MapReduce job usually splits the input data-set into independent chunks which are processed by the map tasks in a completely parallel manner. The framework sorts the outputs of the maps, which are then input to the reduce tasks. Typically both the input and the output of the job are stored in a file-system.
Map Reduce from Prashant Gupta
]]>
17825 7 https://cdn.slidesharecdn.com/ss_thumbnails/mapreduce-170917102008-thumbnail.jpg?width=120&height=120&fit=bounds presentation Black http://activitystrea.ms/schema/1.0/post http://activitystrea.ms/schema/1.0/posted 0
Hadoop File system (HDFS) /slideshow/hadoop-file-system-hdfs/79856624 hdfs-170917101843
HDFS is a Java-based file system that provides scalable and reliable data storage, and it was designed to span large clusters of commodity servers. HDFS has demonstrated production scalability of up to 200 PB of storage and a single cluster of 4500 servers, supporting close to a billion files and blocks.]]>

HDFS is a Java-based file system that provides scalable and reliable data storage, and it was designed to span large clusters of commodity servers. HDFS has demonstrated production scalability of up to 200 PB of storage and a single cluster of 4500 servers, supporting close to a billion files and blocks.]]>
Sun, 17 Sep 2017 10:18:43 GMT /slideshow/hadoop-file-system-hdfs/79856624 PrashantGupta82@slideshare.net(PrashantGupta82) Hadoop File system (HDFS) PrashantGupta82 HDFS is a Java-based file system that provides scalable and reliable data storage, and it was designed to span large clusters of commodity servers. HDFS has demonstrated production scalability of up to 200 PB of storage and a single cluster of 4500 servers, supporting close to a billion files and blocks. <img style="border:1px solid #C3E6D8;float:right;" alt="" src="https://cdn.slidesharecdn.com/ss_thumbnails/hdfs-170917101843-thumbnail.jpg?width=120&amp;height=120&amp;fit=bounds" /><br> HDFS is a Java-based file system that provides scalable and reliable data storage, and it was designed to span large clusters of commodity servers. HDFS has demonstrated production scalability of up to 200 PB of storage and a single cluster of 4500 servers, supporting close to a billion files and blocks.
Hadoop File system (HDFS) from Prashant Gupta
]]>
10054 6 https://cdn.slidesharecdn.com/ss_thumbnails/hdfs-170917101843-thumbnail.jpg?width=120&height=120&fit=bounds presentation Black http://activitystrea.ms/schema/1.0/post http://activitystrea.ms/schema/1.0/posted 0
Apache PIG /slideshow/apache-pig-79856433/79856433 5-170917100629
Apache Pig is a high-level platform for creating programs that runs on Apache Hadoop. The language for this platform is called Pig Latin. Pig can execute its Hadoop jobs in MapReduce, Apache Tez, or Apache Spark.]]>

Apache Pig is a high-level platform for creating programs that runs on Apache Hadoop. The language for this platform is called Pig Latin. Pig can execute its Hadoop jobs in MapReduce, Apache Tez, or Apache Spark.]]>
Sun, 17 Sep 2017 10:06:29 GMT /slideshow/apache-pig-79856433/79856433 PrashantGupta82@slideshare.net(PrashantGupta82) Apache PIG PrashantGupta82 Apache Pig is a high-level platform for creating programs that runs on Apache Hadoop. The language for this platform is called Pig Latin. Pig can execute its Hadoop jobs in MapReduce, Apache Tez, or Apache Spark. <img style="border:1px solid #C3E6D8;float:right;" alt="" src="https://cdn.slidesharecdn.com/ss_thumbnails/5-170917100629-thumbnail.jpg?width=120&amp;height=120&amp;fit=bounds" /><br> Apache Pig is a high-level platform for creating programs that runs on Apache Hadoop. The language for this platform is called Pig Latin. Pig can execute its Hadoop jobs in MapReduce, Apache Tez, or Apache Spark.
Apache PIG from Prashant Gupta
]]>
2182 4 https://cdn.slidesharecdn.com/ss_thumbnails/5-170917100629-thumbnail.jpg?width=120&height=120&fit=bounds presentation Black http://activitystrea.ms/schema/1.0/post http://activitystrea.ms/schema/1.0/posted 0
Map reduce prashant /slideshow/map-reduce-prashant/79600681 mapreduceprashant-170910084845
Mapreduce combiner & partitioner ]]>

Mapreduce combiner & partitioner ]]>
Sun, 10 Sep 2017 08:48:45 GMT /slideshow/map-reduce-prashant/79600681 PrashantGupta82@slideshare.net(PrashantGupta82) Map reduce prashant PrashantGupta82 Mapreduce combiner & partitioner <img style="border:1px solid #C3E6D8;float:right;" alt="" src="https://cdn.slidesharecdn.com/ss_thumbnails/mapreduceprashant-170910084845-thumbnail.jpg?width=120&amp;height=120&amp;fit=bounds" /><br> Mapreduce combiner &amp; partitioner
Map reduce prashant from Prashant Gupta
]]>
801 2 https://cdn.slidesharecdn.com/ss_thumbnails/mapreduceprashant-170910084845-thumbnail.jpg?width=120&height=120&fit=bounds presentation Black http://activitystrea.ms/schema/1.0/post http://activitystrea.ms/schema/1.0/posted 0
Sqoop /slideshow/sqoop-64689627/64689627 sqoop-160804092114
Apache Sqoop efficiently transfers bulk data between Apache Hadoop and structured datastores such as relational databases. Sqoop helps offload certain tasks (such as ETL processing) from the EDW to Hadoop for efficient execution at a much lower cost. Sqoop can also be used to extract data from Hadoop and export it into external structured datastores. Sqoop works with relational databases such as Teradata, Netezza, Oracle, MySQL, Postgres, and HSQLDB]]>

Apache Sqoop efficiently transfers bulk data between Apache Hadoop and structured datastores such as relational databases. Sqoop helps offload certain tasks (such as ETL processing) from the EDW to Hadoop for efficient execution at a much lower cost. Sqoop can also be used to extract data from Hadoop and export it into external structured datastores. Sqoop works with relational databases such as Teradata, Netezza, Oracle, MySQL, Postgres, and HSQLDB]]>
Thu, 04 Aug 2016 09:21:14 GMT /slideshow/sqoop-64689627/64689627 PrashantGupta82@slideshare.net(PrashantGupta82) Sqoop PrashantGupta82 Apache Sqoop efficiently transfers bulk data between Apache Hadoop and structured datastores such as relational databases. Sqoop helps offload certain tasks (such as ETL processing) from the EDW to Hadoop for efficient execution at a much lower cost. Sqoop can also be used to extract data from Hadoop and export it into external structured datastores. Sqoop works with relational databases such as Teradata, Netezza, Oracle, MySQL, Postgres, and HSQLDB <img style="border:1px solid #C3E6D8;float:right;" alt="" src="https://cdn.slidesharecdn.com/ss_thumbnails/sqoop-160804092114-thumbnail.jpg?width=120&amp;height=120&amp;fit=bounds" /><br> Apache Sqoop efficiently transfers bulk data between Apache Hadoop and structured datastores such as relational databases. Sqoop helps offload certain tasks (such as ETL processing) from the EDW to Hadoop for efficient execution at a much lower cost. Sqoop can also be used to extract data from Hadoop and export it into external structured datastores. Sqoop works with relational databases such as Teradata, Netezza, Oracle, MySQL, Postgres, and HSQLDB
Sqoop from Prashant Gupta
]]>
3244 2 https://cdn.slidesharecdn.com/ss_thumbnails/sqoop-160804092114-thumbnail.jpg?width=120&height=120&fit=bounds presentation Black http://activitystrea.ms/schema/1.0/post http://activitystrea.ms/schema/1.0/posted 0
6.hive /slideshow/6hive/63218461 6-160619110045
Apache Hive is a data warehouse infrastructure built on top of Hadoop for providing data summarization, query, and analysis. While developed by Facebook.]]>

Apache Hive is a data warehouse infrastructure built on top of Hadoop for providing data summarization, query, and analysis. While developed by Facebook.]]>
Sun, 19 Jun 2016 11:00:45 GMT /slideshow/6hive/63218461 PrashantGupta82@slideshare.net(PrashantGupta82) 6.hive PrashantGupta82 Apache Hive is a data warehouse infrastructure built on top of Hadoop for providing data summarization, query, and analysis. While developed by Facebook. <img style="border:1px solid #C3E6D8;float:right;" alt="" src="https://cdn.slidesharecdn.com/ss_thumbnails/6-160619110045-thumbnail.jpg?width=120&amp;height=120&amp;fit=bounds" /><br> Apache Hive is a data warehouse infrastructure built on top of Hadoop for providing data summarization, query, and analysis. While developed by Facebook.
6.hive from Prashant Gupta
]]>
1926 11 https://cdn.slidesharecdn.com/ss_thumbnails/6-160619110045-thumbnail.jpg?width=120&height=120&fit=bounds presentation Black http://activitystrea.ms/schema/1.0/post http://activitystrea.ms/schema/1.0/posted 0
Apache HBaseâ„¢ /slideshow/apache-hbase-62392255/62392255 hbaseprashant-160525153503
Apache HBaseâ„¢ is the Hadoop database, a distributed, salable, big data store.Its a column-oriented database management system that runs on top of HDFS. Apache HBase is an open source NoSQL database that provides real-time read/write access to those large data sets. ... HBase is natively integrated with Hadoop and works seamlessly alongside other data access engines through YARN.]]>

Apache HBaseâ„¢ is the Hadoop database, a distributed, salable, big data store.Its a column-oriented database management system that runs on top of HDFS. Apache HBase is an open source NoSQL database that provides real-time read/write access to those large data sets. ... HBase is natively integrated with Hadoop and works seamlessly alongside other data access engines through YARN.]]>
Wed, 25 May 2016 15:35:03 GMT /slideshow/apache-hbase-62392255/62392255 PrashantGupta82@slideshare.net(PrashantGupta82) Apache HBaseâ„¢ PrashantGupta82 Apache HBaseâ„¢ is the Hadoop database, a distributed, salable, big data store.Its a column-oriented database management system that runs on top of HDFS. Apache HBase is an open source NoSQL database that provides real-time read/write access to those large data sets. ... HBase is natively integrated with Hadoop and works seamlessly alongside other data access engines through YARN. <img style="border:1px solid #C3E6D8;float:right;" alt="" src="https://cdn.slidesharecdn.com/ss_thumbnails/hbaseprashant-160525153503-thumbnail.jpg?width=120&amp;height=120&amp;fit=bounds" /><br> Apache HBaseâ„¢ is the Hadoop database, a distributed, salable, big data store.Its a column-oriented database management system that runs on top of HDFS. Apache HBase is an open source NoSQL database that provides real-time read/write access to those large data sets. ... HBase is natively integrated with Hadoop and works seamlessly alongside other data access engines through YARN.
Apache HBaseâ„¢ from Prashant Gupta
]]>
4900 9 https://cdn.slidesharecdn.com/ss_thumbnails/hbaseprashant-160525153503-thumbnail.jpg?width=120&height=120&fit=bounds presentation 000000 http://activitystrea.ms/schema/1.0/post http://activitystrea.ms/schema/1.0/posted 0
Mongodb - NoSql Database /slideshow/mongodb-43492410/43492410 nosqlmongodb-150113235205-conversion-gate02
MongoDB is an open-source document database, and the leading NoSQL database. Written in C++. MongoDB has official drivers for a variety of popular programming languages and development environments. There are also a large number of unofficial or community-supported drivers for other programming languages and frameworks.]]>

MongoDB is an open-source document database, and the leading NoSQL database. Written in C++. MongoDB has official drivers for a variety of popular programming languages and development environments. There are also a large number of unofficial or community-supported drivers for other programming languages and frameworks.]]>
Tue, 13 Jan 2015 23:52:05 GMT /slideshow/mongodb-43492410/43492410 PrashantGupta82@slideshare.net(PrashantGupta82) Mongodb - NoSql Database PrashantGupta82 MongoDB is an open-source document database, and the leading NoSQL database. Written in C++. MongoDB has official drivers for a variety of popular programming languages and development environments. There are also a large number of unofficial or community-supported drivers for other programming languages and frameworks. <img style="border:1px solid #C3E6D8;float:right;" alt="" src="https://cdn.slidesharecdn.com/ss_thumbnails/nosqlmongodb-150113235205-conversion-gate02-thumbnail.jpg?width=120&amp;height=120&amp;fit=bounds" /><br> MongoDB is an open-source document database, and the leading NoSQL database. Written in C++. MongoDB has official drivers for a variety of popular programming languages and development environments. There are also a large number of unofficial or community-supported drivers for other programming languages and frameworks.
Mongodb - NoSql Database from Prashant Gupta
]]>
3211 2 https://cdn.slidesharecdn.com/ss_thumbnails/nosqlmongodb-150113235205-conversion-gate02-thumbnail.jpg?width=120&height=120&fit=bounds presentation Black http://activitystrea.ms/schema/1.0/post http://activitystrea.ms/schema/1.0/posted 0
Sonar Tool - JAVA code analysis /slideshow/sonar-tool-java-code-analysis/43492358 sonarcodeanalysis-150113234820-conversion-gate01
SonarQube is an open platform to manage code quality. It has got a very efficient way of navigating, a balance between high-level view, dashboard, TimeMachine and defect hunting tools. SonarQube tool is a web-based application. Rules, alerts, thresholds, exclusions, settings… can be configured online.]]>

SonarQube is an open platform to manage code quality. It has got a very efficient way of navigating, a balance between high-level view, dashboard, TimeMachine and defect hunting tools. SonarQube tool is a web-based application. Rules, alerts, thresholds, exclusions, settings… can be configured online.]]>
Tue, 13 Jan 2015 23:48:19 GMT /slideshow/sonar-tool-java-code-analysis/43492358 PrashantGupta82@slideshare.net(PrashantGupta82) Sonar Tool - JAVA code analysis PrashantGupta82 SonarQube is an open platform to manage code quality. It has got a very efficient way of navigating, a balance between high-level view, dashboard, TimeMachine and defect hunting tools. SonarQube tool is a web-based application. Rules, alerts, thresholds, exclusions, settings… can be configured online. <img style="border:1px solid #C3E6D8;float:right;" alt="" src="https://cdn.slidesharecdn.com/ss_thumbnails/sonarcodeanalysis-150113234820-conversion-gate01-thumbnail.jpg?width=120&amp;height=120&amp;fit=bounds" /><br> SonarQube is an open platform to manage code quality. It has got a very efficient way of navigating, a balance between high-level view, dashboard, TimeMachine and defect hunting tools. SonarQube tool is a web-based application. Rules, alerts, thresholds, exclusions, settings… can be configured online.
Sonar Tool - JAVA code analysis from Prashant Gupta
]]>
2198 3 https://cdn.slidesharecdn.com/ss_thumbnails/sonarcodeanalysis-150113234820-conversion-gate01-thumbnail.jpg?width=120&height=120&fit=bounds presentation Black http://activitystrea.ms/schema/1.0/post http://activitystrea.ms/schema/1.0/posted 0
https://cdn.slidesharecdn.com/profile-photo-PrashantGupta82-48x48.jpg?cb=1627732818 'Software Savant' with over 6.5 years of experience and dedication in development/maintenance/production support projects. Diverse background includes analyzing, designing, developing and deploying and Telecom and Banking domains. Superior record of delivering simultaneous large-scale projects on time. Excellent interpersonal/communication skills coupled with technical and organizational skills to keep all the stake holders happy. Extremely passionate about emerging technologies & seeks to constantly innovate https://cdn.slidesharecdn.com/ss_thumbnails/spark2-171008044051-thumbnail.jpg?width=320&height=320&fit=bounds slideshow/spark-core-80569932/80569932 Spark core https://cdn.slidesharecdn.com/ss_thumbnails/sparksql-171008043944-thumbnail.jpg?width=320&height=320&fit=bounds slideshow/spark-sql-and-dataframe/80569923 Spark Sql and DataFrame https://cdn.slidesharecdn.com/ss_thumbnails/mapreduce-170917102008-thumbnail.jpg?width=320&height=320&fit=bounds slideshow/map-reduce-79856653/79856653 Map Reduce