Category Java 9
Java 9 Modular Programming Concept is a powerful feature which is introduced in JDK 1.9. Modularity specifies the interrelation and intercommunication between the parts that comprise a softwaresystem. Modular programming...
Why do we need Java 9 in Real word. What it offers that you should look into and move your application to production without any second thought, here I have...
Java 9 Programming By Example Peter Verhas is a book written for Java 9 developers, indeed it is targeted for java programmers with lot of supporting examples and build information....
Category Book Review
Hadoop 3.0 Security by Ben and Joey: Protecting Your Big Data Platform is an excellent, practical, well-written book which describes the Apache Hadoop and the numerous security features within Apache...
All these terms “Blockchain Bitcoin and Decentralized Applications” are now synonymous words and I decided to read the book Decentralized Application by Siraj Raval and share my experience with all...
Java 9 Programming By Example Peter Verhas is a book written for Java 9 developers, indeed it is targeted for java programmers with lot of supporting examples and build information....
Category Blockchain
All these terms “Blockchain Bitcoin and Decentralized Applications” are now synonymous words and I decided to read the book Decentralized Application by Siraj Raval and share my experience with all...
Category Software Architecture
Java programming language has always been criticised for poor performance. However when it comes to enterprise applications, Java has been the first choice for architects. Don’t they sounds conflicting, actually...
All these terms “Blockchain Bitcoin and Decentralized Applications” are now synonymous words and I decided to read the book Decentralized Application by Siraj Raval and share my experience with all...
Category Software Development
This is a very tricky question and asked in certification exam as well as in many interviews and you would find a lot of discussion about “Concept of Sprint Zero”....
SCRUM Master is a very crucial role in Agile SCRUM software development and best way to understand the role and responsibility of a scrum master is educate yourself by reading...
Many software engineering, IT professionals and IT business analyst, who aspire to be a “SCRUM Master”, look around to get an answer for “How to Pass Professional Scrum Master” or...
Category Agile SCRUM
What is the best way to measure progress in agile project? Most of the people follow burn down chart and JIRA agile dashboard helps you to add it. Measuring Progress...
This is a very tricky question and asked in certification exam as well as in many interviews and you would find a lot of discussion about “Concept of Sprint Zero”....
SCRUM Master is a very crucial role in Agile SCRUM software development and best way to understand the role and responsibility of a scrum master is educate yourself by reading...
Many software engineering, IT professionals and IT business analyst, who aspire to be a “SCRUM Master”, look around to get an answer for “How to Pass Professional Scrum Master” or...
Java programming language has always been criticised for poor performance. However when it comes to enterprise applications, Java has been the first choice for architects. Don’t they sounds conflicting, actually...
Category JIRA Agile
What is the best way to measure progress in agile project? Most of the people follow burn down chart and JIRA agile dashboard helps you to add it. Measuring Progress...
Builing JIRA Dashboard is such an easy task and anybody can be Naive to Nijna in just 1 hours. If you are part of Agile Team, you must have a...
An Agile SCRUM project has 3 roles named as Product Owner (PO), Scrum Master (SM) and Development Team. This “ JIRA Agile General Purpose Guide” aiming to explain how to...
SCRUM Master and Product Owner Same Person kind of arrangement is discouraged by SCRUM if you are a follower of SCRUM.org, but it is technically possible. SCRUM Master has much...
Most of the people who has used JIRA for defect tracking purpose think that JIRA is just for logging defect and tracking them via defect life cycle. JIRA is a...
Category Java Certification 8
There are 900,000 Java developers and 30,000java developer get Oracle Java Certification Examsevery year. This clearly indicates that how popular this programming language is. The java community isusing the knowledge,...
Java Certification Test Paper on Operator &Statements :Every line in java is an statement and it may use an operator, how to use it effectively and wisely depends on developer’s...
Java Certification Test Paper on Java Building Block topic expect the candidates to be aware of building blocks of java. This article not only covers the exam essential but also...
Oracle Java Certification is one of the most regarded certification for java developers. It is important to understand the Oracle Java Certification Path before you make a plan for the...
In this Java Certification Tips Tricks, I have provided a very simple and powerful techniques how to prepare for OCA certification exam with very minimal preparation. If you are already...
You need to recognise three types of exceptions for the OCA exam: run-time exceptions,checked exceptions, and errors. Well look at common examples of each type. For theexam, youll need to...
Exception Handling Java 8 Certification (or Catching Various Types of Exceptions) is an key aspect of application development when working with Java.So far, you have been catching only one type...
Try Catch Block (on Try Catch Block Java 8 Certification Exam) is most widely used statement when it comes to manage exception. Atry statement must have catch and/or finally.Having both...
Exceptions Java 8 Certification(OCA Exam) covers only thebasics of working with exceptions. Understanding the role of exceptions and type of exceptions are very important for Java Certification. You must read...
Class Design Java 8 Certificationis another important chapter for when appearing for OCA (Oracle Certified Associate Exam 1Z0-808), (or for Java 8 Certification). At its core, proper Java class design...
Core API Java 8 Certification(or in short we can call it: Core Java API)is another important chapter for when appearing for OCA (Oracle Certified Associate Exam 1Z0-808), (or for Java...
Operators and Statementsis the next important chapter for when appearing for OCA (Oracle Certified Associate Exam 1Z0-808), (or for Java 8 Certification). You must understand the Operator and Statementfundamental. This...
Java Building Blocksis the first important chapter for When appearing for OCA (Oracle Certified Associate Exam 1Z0-808), (or for Java 8 Certification). You must understand the Java Basics & Java...
OCJP Practice Papers Java Concurrency include following topics Create worker threads using Runnable, Callable and use an ExecutorService toconcurrently execute tasks Identify potential threading problems among deadlock, starvation, livelock, andrace...
OCJP Practice Papers Java File I/O NIO.2 include following topics Use Path interface to operate on file and directory paths Use Files class to check, read, delete, copy, move, manage...
OCJP Practice Papers Selected Classes from Java API include following topics Manipulate data using the StringBuilder class and its methods Create and manipulate Strings Create and manipulate calendar data using...
OCJP Practice Papers Java Handling Exceptions include following topics Differentiate among checked exceptions, unchecked exceptions, and Errors Create a try-catch block and determine how exceptions alter normal program flow Describe...
OCJP Practice Papers Java Stream API include following topics Describe inheritance and its benefits Develop code that makes use of polymorphism; develop code that overrides methods; differentiate between the type...
OCJP Practice Papers Java Stream API include following topics Create methods with arguments and return values; including overloaded methods Apply the static keyword to methods and fields Create and overload...
OCJP Practice Papers Java I/O Fundamentals include following topics Read and write data from the console Use BufferedReader, BufferedWriter, File, FileReader, FileWriter, FileInputStream, FileOutputStream, ObjectOutputStream, ObjectInputStream, andPrintWriter in the java.io...
OCJP Practice Papers Date/Time API include following topics Create and manage date-based and time-based events including a combination ofdate and time into a single object using LocalDate, LocalTime, LocalDateTime,Instant, Period,...
OCJP Practice Papers Java Stream API include following topics Use try-catch and throw statements Use catch, multi-catch, and finally clauses Use Autoclose resources with a try-with-resources statement Create custom exceptions...
OCJP Practice Papers Java Stream API include following topics Develop code to extract data from an object using peek() and map() methods including primitive versions of the map() method Search...
OCJP Practice Papers Java Class Design include following topics Implement encapsulation Implement inheritance including visibility modifiers and composition Implement polymorphism Override hashCode, equals, and toString methods from Object class Create...
OCJP Practice Papers Using Loop Constructs include following topics Create and use a generic class Create and use ArrayList, TreeSet, TreeMap, and ArrayDeque objects Use java.util.Comparator and java.lang.Comparable interfaces Collections...
OCJP Practice Papers Using Loop Constructs include following topics Create and use a generic class Create and use ArrayList, TreeSet, TreeMap, and ArrayDeque objects Use java.util.Comparator and java.lang.Comparable interfaces Collections...
OCJP Practice Papers Using Loop Constructs include following topics Create and use while loops Create and use for loops including the enhanced for loop Create and use do/while loops Compare...
OCJP Practice Papers Creating and Using Arrays include following topics Declare, instantiate, initialize and use a one-dimensional array Declare, instantiate, initialize and use multi-dimensional arrays See the complete syllabus for...
OCJP Practice Papers Advanced Java Class Design Include following topics Develop code that uses abstract classes and methods Develop code that uses the final keyword Create inner classes including static...
OCJP Practice Papers - Operators and Decision Constructs covers following topic under this section Using Operators and Decision Constructs Use Java operators; use parentheses to override operator precedence Test equality...
OCJP Practice Papers - Working with Java Data Types
OCJP Practice Papers - Java Basics : Covers all the basic questions
Java 8 Certification or OCJP Certification Syllabus (Oracle Certified Java Profession) has two certification exams. The 1st certification exam is called OCA (Oracle Certified Associate 1Z0-808). This java certification is...
Category Java Interview
If you are reading my title “Java Interview Question Tips For Entry Level Developers”, I can consider that you are somewhere around 0-1 year of experience java developer. You might...
Category Data Engineering Concepts
It seems like a trial task, but standars naming convention are important. Good names for tables, columns and other database objects are particularly important for ad hoc users of the...
What is data lineage and why it is important. Data lineage is nothing but its origins and transformation that data goes through with time. Data lineage can also be expressed...
Data Lineage and Data Provenance are not the same thing. Many data engineer and architect use them interchangible but they are two different concept and has its separate meaning.
CAP Theorem was proposed by Eric Brewer (professor of computer science at the University of California, Berkeley, and vice president of infrastructure at Google) in year 2000. Eric found that...
Category Java 8
Java Certification Test Paper on Operator &Statements :Every line in java is an statement and it may use an operator, how to use it effectively and wisely depends on developer’s...
Java Building Blocksis the first important chapter for When appearing for OCA (Oracle Certified Associate Exam 1Z0-808), (or for Java 8 Certification). You must understand the Java Basics & Java...
Category Cloudera Impala
The Beginners Impala Tutorial covers key concepts of in-memory computation technology called Impala. It is developed by Cloudera. MapReduce based frameworks like Hive is slow due to excessive I/O operations....
Category Apache Spark
How to perform minus operation on a date type or timestamp time.
Apache Spark is an exciting technology that is rapidly superseding Hadoops MapReduce as the preferred big data processing platform. Hadoop is an open source, distributed, Java computation framework consisting of...
Immutability and RDD Interface in Spark are key concepts and it must be understood in detail.Spark defines an RDD interface with the properties that each type of RDD mustimplement. These...
Spark In-Memory Persistence and Memory Management must be understood by engineering teams.Sparks performance advantage over MapReduce is greatest in use cases involvingrepeated computations. Much of this performance increase is due...
Spark Model of Parallel Computing and sometimes also called RDD is an important API.Spark Model of Parallel Computing internally uses RDD and part of Spark Core library.
Why you should be worried about how Apache Spark works? To get the most out of Spark, it is important to understand some of the principlesused to design Spark and,...
Category Chef
One of the most well-known differences between managing UNIX-like systems and Windows systems is the Windows Registry. Chef has resources for creating, modifying, and deleting Windows Registry keys. Beware that...
A large number of managed systems require configuration of software that is outside the scope of the built-in Windows roles and features. Chef has a very handy resource for installing...
Similar to Linux script resources for bash, ruby, and so on, Chef can execute arbitrarily-defined Windows batch scripts through the command interpreter. When these resources are used, Chef compiles the...
While using Chef for Windows, there are multiple backends for the Windows feature resourceDISM and servermanagercmd. Each one has a specific Ruby class that will be used based on the...
Similar to how Linux distributions have package management tools and a repository of packages, Windows has long had built-in packages that come with the OS. Both desktop and server releases...
When managing Windows with Chef, there are some Windows-specific resources that are available to you as part of the Windows stack. This section covers those resources that are specific to...
Chef is a very popular infrastructure automation framework. It is also getting popular in windows based environment. Installing Chef Client on Windows is sometime a challenge for developer. In order...
Chef and Microsoft Windows are getting very popular. As chef practitioner, you must learn how chef works on Microsoft Windows. As you know that the client-side components of Chef are...
Category Apache NiFi
Apache NiFi is a software project from the Apache Software Foundation designed to automate the flow of data between software systems (file system, RDBMS, APIs etc in and out) ....
Category PySpark
What is PySpark (is it pySpark or PySpark)? Is it a new programming language or just another library? Or it is very different from Apache Spark? In recnet past, PySpark...
This guide on PySpark Installation on Windows 10 will provide you a step by step instruction to make Spark/Pyspark running on your local windows machine. Most of us who are...
PySpark on Windows can be installed using two different ways. Since Spark is a distributed compute engine, it also works stand alone. Most of the developer who are familiar with...
About PySpark Local Installation Can I install PySpark in Windows 10? Yes, PySpark can be installed in Windows 10 or even earlier version of it, refer complete guide here. General...
In this PySpark Tutorial, we will understand why PySpark is becoming popular among data engineers and data scientist. This PySpark Tutorial will also highlight the key limilation of PySpark over...
Category Apache Hive
Rearranging columns in PySpark or Spark (Scala) data frame should not be a difficult job. PySpark or Spark/Scala provides a lot of convenient APIs to arrange the columns to meet...
Apache Hive does not have direct standard UDF for transposing rows into columns. Transpose & Pivot in Hive Query can be achieved using multi-stage process. You can use collect_list() or...
Apache Hive Vectorization was introduced newly in Apache Hive to improve query performance. By default, the Apache Hive query execution engine processes one row of a table at a time....
Apache Hive Release 3.1.1 is the version which is compatible with Hadoop 3.x.y and fixes 4 bugs and one new Feature Apache Hive Release 3.1.1 Release Note Following Bug Fixes...
Apache Hive development has shifted from the original Hive server (HiveServer1) to the new server (HiveServer2), and hence users and developers need to move to the new access tool. However,...
Apache Hive Cheat Sheet is a summary of all functions and syntax for big data engineers and developers reference. It is divided into 5 parts. Apache Hive Cheat Sheet -...
As big data engineer, you must know the apachehive best practices.As you know Apache Hive is not an RDBMS, but it pretends to be one most of the time. It...
Apache Hive Analytical Functionsavailable since Hive 0.11.0, are a special group of functions that scan the multiple input rows to compute each output value. Apache Hive Analytical Functions are usually...
Category Apache Sqoop
This guide helps you in downloading and installing apache sqoop. Apache Sqoop supports the Linux operating system, and there are several installation options. One option is the source tarball that...
In this article, Apache Sqoop Introduction, we will primarily discuss why this tool exists. Apache sqoopis part of Hadoop Core project or part of Hadoop Ecosystem project.
Category Hadoop 3.0
Many spark users who are using Hadoop as storage under the Spark computation is asking if Hadoop 3.0 vs Spark 2.x compatible or not. Spark 2.2.1 was released the 1st...
Hadoop 3.0 vs Hadoop 2.0 : Hadoop 3.0.0 GA (General Availability) is released on 13-Dec-2017. Everybody wants to know what it brings into the table for developer, administrator and enterprise...
Hadoop 3.0 Security by Ben and Joey: Protecting Your Big Data Platform is an excellent, practical, well-written book which describes the Apache Hadoop and the numerous security features within Apache...
Hadoop 3.0 Roadmap : Latest version of Hadoop is already out and everybody is excited about itsfeatures. Hadoop 3.0 is a major release after Hadoop 2.9 which was released in...
Hadoop 3.0 Release Dateis 13-Dec-2017 and General Availability (GA) binaries can be downloaded from its official website. General availability (GA) marks a point of quality and stability for the release...
Why do you need Hadoop 3.0 new features and enhancement and what is driving this change. The key driving force behind Hadoop 3.0 that there are lot of content in...
Hadoop 3.0 or Bigdata jobs are in demand and in Hadoop 3.0 Interview Question article covers almost all the important topic including the reference link to other tutorials.
Apache Hadoop 3.0 Installation on Windows is a short and practical guide for bigdata engineersto get their hands dirty. Since Hadoop 3.0 is not yet available with Cloudera CDH 6.x...
This blog, Apache Hadoop 3.0 Installation, will assist you install and verify your pseudo-distributed, single-node & distributed instance in UNIX box (RHLE or Ubuntu). Hadoop 3.0 needs java 1.8 and...
Hadoop 3.0 Hortonworks is an obvious question after you have seen Hadoop 3.0 new feature and enhancement list. At the time of writing this blog, HDP was having 2.6.4 supporting...
Hadoop 3.0 GPU : Hadoop is still behind high performance capacity due to CPUs limited parallelism, though. GPU (Graphical Processing Unit) accelerated computing involves the use of a GPU together...
This deep dive article on “Hadoop 3.0 Erasure coding explained” will highlight how the erasure coding will help reducing 50% of storage overhead cost. The storage component (HDFS) of Hadoop...
The official Apache Hadoop 3.0 Download was made availableDec 2017. The Hadoop 3.0 is a feature packed release with lots of new feature and enhancements.Since Hadoop 3.0 is not yet...
Hadoop 3.0 Docker : The docker enables users to bundle an application together with its preferred execution environment. In this article, we will talk about Hadoop and docker together. What...
Hadoop 3.0 Cloudera (or CDH 6.x)is an obvious question after you have seen Hadoop 3.0 new feature and enhancement list. At the time of writing this blog, CDH was having...
Hadoop, the most popular open-source distributed framework has arrived with a new release 3.x. It brings promisingfeatures and enhancements, but here we will demystify the Hadoop 3.0 Architecture in detail....
Category Hadoop
Hadoop 3.0 or Bigdata jobs are in demand and in Hadoop 3.0 Interview Question article covers almost all the important topic including the reference link to other tutorials.
Category Unix
While working with UNIX (Ubuntu/CentOS/MAC) terminals, it is not very convenient to move your cursors back and forth and it makes our overall productivity low. Since these terminals do not support...
This short article talks and compare UNIX kernel shells, which many technical folks are confused of. The Unix operating system used a shell program called theBourne Shell. Then, slowly, many...
Many times you get access to a Unix or Linux box via terminal. Before start using the terminal, you may want to know the Unix flavor. This article will help...
Category Snowflake Data Warehouse
How to Load, Query And Proces JSON Data Into Snowflake
Before we process and chrunch data, it has to be loaded and it has to be loaded effectively and efficiently. This article talks about all the best practices for loading...
This blogs explain how to handle date and time formats while loading data into snowflake tables from csv files and how to make sure that milli and nano second values...
This blogs explain how to handling null fields in your data file, be it CSV or delimiter separated files and load the data set effectively in your snowflake tables.
This blogs explain how to run validation process while loading small or large CSV files into snowflake. This can be done using a parameter and it has certain limitations and...
Data duplication is a common issue in many data engineering work and when it happens, it becomes a costly matter. So this blog describe how to de-duplicate data when a...
How to load data special characters available in your data file whicch either conflicts with your file delimiter or row delimiter or stops your copy command and eventually stops data...
How to load data into snowflake
Snowflake Python Connector Installation, Configuration, Example & Data Loading To Stage
Snowflake ODBC Driver Installation, Configuration, Example & Microsoft Excel Data Loading
How To Use Snowflake Java JDBC Type-4 Driver to build utilities
What is information Schema or How to get information schema in snowflake
Type of views (standard, secure & materialized) in Snowflake, their SQL construct, their limitations, cost implication with materialized views and best practices around it.
How to use standard, secure & materialized views in Snowflake, their SQL construct, their limitations, cost implication with materialized views and best practices around it.
How to use User Defined Functions in Snowflake using SQL Sript, JavaSript & Java Programming Language.
How to use strict property in Snowflake UDFs
How to use not-null property in Snowflake UDFs
How to use strict property in Snowflake UDFs
How to use secure property in Snowflake UDFs
How to use function overloading in Snowflake UDFs
How to use language property in Snowflake UDFs
How to use strict property in Snowflake UDFs
How to use function overloading in Snowflake UDFs
How to use snowflake stored procedure using javascript
Role & Access Control is most powerful concept in snowflake and every snowflake developer must understand it.
Watch E2E Snowflake ETL Demo
Watch E2E Snowflake ETL Demo
Snowflake ETL Using Pipe, Stream & Task
Watch E2E Snowflake ETL Demo
Watch E2E Snowflake ETL Demo
Watch E2E Snowflake ETL Demo
Snowflake ETL Using Pipe, Stream & Task
Snowflake Time Travel Partition
Introduction Data pipelines automate many of the manual steps involved in transforming and optimizing continuous data loads. Frequently, the “raw” data is first loaded temporarily into a staging table (stage...
Time Travel is most powerful feature introduced by snowflake, and it will be discussed in detail in this article.
Snowflake Micro Partition
Complete Guide on Snowflake External Table
Snowflake Unique Features (Key Unique Features)
Continous Data Loading is trickier thing in snowflake and how to load it wihout any external stage is evey trickier. This blog will focus how to load data using a...
Snowflake Container Hierarchy concept is very important and not understood by many developer. This blog focuses with table creation, be it standard or external or transient or temporary. It also...
What does is_role_in_session() do?
What does last_query_id() do?
What does current_ip_address() do?
What are snowflake date and time context functions ?
Why every snowflake developer must know about task object in snowflake.
Why you should consider Snowflake Cloud Data Warehouse and what does it bring to the table.
What is the context function in the Snowflake?
What is Stream (Object) in Snowflake?
Snowflake Data Warehouse Glossary talks about all the important keywords which a developer must know. Snowflake cloud data platform is cloud-native, faster, easier to use, and far more flexible than...
Snowflake is a cloud-native and easy to use virtual data warehouse system. Since it is built on the top of the cloud-native platform, traditional best practices do not applicable anymore....
Category /apache-spark
This write up talks about Apache Spark 3.0 features and improvements. Apache Spark 3.0 is a major release and currently available in preview mode. Release 3.0.0 is a major release...
Category release-3.0.0/
This write up talks about Apache Spark 3.0 features and improvements. Apache Spark 3.0 is a major release and currently available in preview mode. Release 3.0.0 is a major release...
Category Apache Spark 3
scala spark spark-three datasource-v2-spark-three Spark 3.0 is a major release of Apache Spark framework. It’s been in preview from last December and going to have a stable release very soon....
Apache Spark 3.0 is released and available for testing in preview mode. The release was done on 2019-Nov-08 and it was announed via twiter. The preview mode is lauched to...
Category Snowflake SnowPro Certification
Snowflake Data Sharing Introduction (Snowflake Certification) Data Sharing is one of the key differentiator in Snowflake and certification exam will ask 4-5 questions on this subject area. Data Sharing feature...
SnowPipe Introduction (Snowflake Certification) Data Loading & Data Ingestion is one of the key activity in any data warehouse system and that’s why Snowflake SnowPro Certification exam ask many questions...
Category Azure Data Engineering
Category User Defined Functions
How to use strict property in Snowflake UDFs
How to use not-null property in Snowflake UDFs
How to use strict property in Snowflake UDFs
Category Snowflake Data Loading
Before we process and chrunch data, it has to be loaded and it has to be loaded effectively and efficiently. This article talks about all the best practices for loading...
This blogs explain how to handle date and time formats while loading data into snowflake tables from csv files and how to make sure that milli and nano second values...
This blogs explain how to handling null fields in your data file, be it CSV or delimiter separated files and load the data set effectively in your snowflake tables.
This blogs explain how to run validation process while loading small or large CSV files into snowflake. This can be done using a parameter and it has certain limitations and...
Category CSV Data Loading In Snowflake
This blogs explain how to handle date and time formats while loading data into snowflake tables from csv files and how to make sure that milli and nano second values...
This blogs explain how to handling null fields in your data file, be it CSV or delimiter separated files and load the data set effectively in your snowflake tables.
This blogs explain how to run validation process while loading small or large CSV files into snowflake. This can be done using a parameter and it has certain limitations and...
Category Snowflake Best Practice
Before we process and chrunch data, it has to be loaded and it has to be loaded effectively and efficiently. This article talks about all the best practices for loading...
Category JSON Snowflake
How to Load, Query And Proces JSON Data Into Snowflake